<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.extremist.software/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=198.27.140.180</id>
	<title>Noisebridge - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.extremist.software/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=198.27.140.180"/>
	<link rel="alternate" type="text/html" href="https://wiki.extremist.software/wiki/Special:Contributions/198.27.140.180"/>
	<updated>2026-04-09T09:16:25Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.39.13</generator>
	<entry>
		<id>https://wiki.extremist.software/index.php?title=Conflict_Resolution&amp;diff=59750</id>
		<title>Conflict Resolution</title>
		<link rel="alternate" type="text/html" href="https://wiki.extremist.software/index.php?title=Conflict_Resolution&amp;diff=59750"/>
		<updated>2017-07-22T02:41:11Z</updated>

		<summary type="html">&lt;p&gt;198.27.140.180: /* Mindfulness towards Escalation: */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==Suggestions for if you are having problematic interactions within the Noisebridge community==&lt;br /&gt;
&lt;br /&gt;
===Personal Confrontation:===&lt;br /&gt;
Is someone bothering you? Talk to them about it -- and be excellent while doing so. &amp;lt;br/&amp;gt;&lt;br /&gt;
This is a pretty important step, and it usually has the desired effect. It should not be skipped if at all possible.&lt;br /&gt;
&lt;br /&gt;
===Get Support:===&lt;br /&gt;
Did that not work?  Or are you afraid to approach the other person?  Ask someone else around the space who you like and/or trust.  Maybe they can come along with you to talk to them, or talk to them as your proxy.&lt;br /&gt;
&lt;br /&gt;
===Mediation:===&lt;br /&gt;
Did that not work?  We have a [[Mediation]] page, where people can sign up to act as mediators. You should ask one of the people on it to help you mediate your conflict.  They can actively mediate a discussion between you and the person with whom you are having conflict, or, if you prefer, the mediator can talk to that individual as your proxy.&lt;br /&gt;
&lt;br /&gt;
===Mindfulness towards Escalation:===&lt;br /&gt;
If it seems appropriate, after talking with the original parties, the mediator (and indeed everyone involved) should start to tactfully ask around and find out if this is an isolated conflict or a more generalized problem in the community. Most personal problems at Noisebridge can be resolved through a series of calm one-on-one talks, and almost all of the rest can be solved by a series of mediated discussions. If mediation is unsuccessful, or if what is going on appears to be part of a larger pattern, the mediator may suggest that you bring your problem to a Safe Space Working Group for discussion. See [[Deescalation|here]] for more info on deescalation.&lt;br /&gt;
&lt;br /&gt;
===Advocate:===&lt;br /&gt;
Discussing personal conflicts at the larger group level is not really considered all that excellent.  On the other hand, a small supportive group environment more specifically committed to calm discussion and de-escalation &#039;&#039;&#039;can&#039;&#039;&#039; help defuse a problematic situation.  If the parties involved cannot reach a resolution by talking with each other, or with the help of a [[Mediation|mediator]], the mediator can suggest calling a meeting of the Safe Space Working Group to involve other people to help resolve the conflict. If you try to follow these suggestions, that would be totally excellent.&lt;br /&gt;
&lt;br /&gt;
Before a problem with an individual is brought to the level of calling a meeting of the Safe Space Working Group, someone must step forward to act as an advocate for the individual, even if that individual happens to be widely disliked. It is all too easy for conflict to make people act in ways that they later regret.  There are sufficient people around the Space who are willing to act as advocates at the group level (see list of mediator volunteers on the [[Mediation]] wiki page).&lt;br /&gt;
&lt;br /&gt;
===Reporting Misbehavior===&lt;br /&gt;
If you wish to report harassment anonymously or privately, you can send a message to secretary@noisebridge.net (or contact one of the people who have volunteered as a mediator on the [[Mediation|Mediation]] page).&lt;br /&gt;
&lt;br /&gt;
{{ManualPage}}&lt;/div&gt;</summary>
		<author><name>198.27.140.180</name></author>
	</entry>
	<entry>
		<id>https://wiki.extremist.software/index.php?title=DreamTeam/Reading&amp;diff=59523</id>
		<title>DreamTeam/Reading</title>
		<link rel="alternate" type="text/html" href="https://wiki.extremist.software/index.php?title=DreamTeam/Reading&amp;diff=59523"/>
		<updated>2017-07-07T05:11:21Z</updated>

		<summary type="html">&lt;p&gt;198.27.140.180: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;(note wiki contains some useful clues re previous neuro research at Noisebridge ... For example, the [[Analog_EEG_Amp]] page describes some project ideas and work done by others here in 2012) &lt;br /&gt;
&lt;br /&gt;
https://noisebridge.net/wiki/NBDSM -- noiseBridge Deepnet and Statistical Mechanics -- first meetup @ noisebridge 7/6/17 at 7PM&lt;br /&gt;
&lt;br /&gt;
https://metacademy.org/&lt;br /&gt;
-- machine learning knowledge graph&lt;br /&gt;
&lt;br /&gt;
http://www.thetalkingmachines.com/ -- podcast&lt;br /&gt;
&lt;br /&gt;
https://karpathy.github.io/2015/05/21/rnn-effectiveness/&lt;br /&gt;
&lt;br /&gt;
http://alexandre.barachant.org/papers/&lt;br /&gt;
&lt;br /&gt;
http://ncs.ethz.ch/publications -- neuromorphic cognitive systems&lt;br /&gt;
&lt;br /&gt;
== Game Theory ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/abs/1707.01068v1 -  Maintaining cooperation in complex social dilemmas using deep reinforcement learning&lt;br /&gt;
&lt;br /&gt;
== Questions of Physics and Free Will ==&lt;br /&gt;
&lt;br /&gt;
http://www.scottaaronson.com/papers/giqtm3.pdf - The Ghost in the Quantum Turing Machine&lt;br /&gt;
&lt;br /&gt;
== CNN ==&lt;br /&gt;
&lt;br /&gt;
https://adeshpande3.github.io/adeshpande3.github.io/A-Beginner&#039;s-Guide-To-Understanding-Convolutional-Neural-Networks/ - &amp;quot;A Beginner&#039;s Guide To Understanding Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://adeshpande3.github.io/A-Beginner&#039;s-Guide-To-Understanding-Convolutional-Neural-Networks-Part-2/ - &amp;quot;A Beginner&#039;s Guide To Understanding Convolutional Neural Networks Part 2&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://scs.ryerson.ca/~aharley/vis/harley_vis_isvc15.pdf -- &amp;quot;An Interactive Node-Link Visualization&lt;br /&gt;
of Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://white.stanford.edu/teach/index.php/An_Introduction_to_Convolutional_Neural_Networks&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Dosovitskiy_Learning_to_Generate_2015_CVPR_paper.pdf -- &amp;quot;Learning to Generate Chairs With Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.pamitc.org/cvpr15/files/lecun-20150610-cvpr-keynote.pdf&lt;br /&gt;
-- &amp;quot;What&#039;s Wrong With Deep Learning?&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Mind-Body Relations ==&lt;br /&gt;
&lt;br /&gt;
http://www.pnas.org/content/111/20/7379.full.pdf -- &amp;quot;Voluntary activation of the sympathetic nervous system and attenuation of the innate immune response in humans&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Math ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1311.1090.pdf -- &amp;quot;Polyhedrons and Perceptrons Are Functionally Equivalent&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Bayesian Inference ==&lt;br /&gt;
&lt;br /&gt;
https://noisebridge.net/images/e/ef/Perception_is_in_the_Details12.pdf --&lt;br /&gt;
&amp;quot;Perception is in the Details: A Predictive Coding Account of the Psychedelic Phenomenon&amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
http://rsif.royalsocietypublishing.org/content/10/86/20130475 --&lt;br /&gt;
&amp;quot;Life as we know it&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://jmlr.csail.mit.edu/proceedings/papers/v31/wang13b.pdf --&lt;br /&gt;
&amp;quot;Collapsed Variational Bayesian Inference for Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.datalab.uci.edu/papers/nips06_cvb.pdf --&lt;br /&gt;
&amp;quot;A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cnbc.cmu.edu/~tai/papers/lee_mumford_josa.pdf --&lt;br /&gt;
&amp;quot;Hierarchical Bayesian inference in the visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.researchgate.net/profile/Til_Bergmann/publication/262423308_Temporal_coding_organized_by_coupled_alpha_and_gamma_oscillations_prioritize_visual_processing/links/0deec537d1bfda474c000000/Temporal-coding-organized-by-coupled-alpha-and-gamma-oscillations-prioritize-visual-processing.pdf --&lt;br /&gt;
&amp;quot;Temporal coding organized by coupled alpha and gamma oscillations prioritize visual processing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cell.com/neuron/pdf/S0896-6273(15)00823-5.pdf --&lt;br /&gt;
&amp;quot;Rhythms for Cognition: Communication through Coherence&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.biorxiv.org/content/biorxiv/early/2014/05/06/004804.full.pdf --&lt;br /&gt;
&amp;quot;Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Speech Recognition ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.00694v1 -- &amp;quot;ESE: Efficient Speech Recognition Engine with Compressed LSTM on FPGA&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Sound Classification ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1608.04363v2 -- &amp;quot;Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1605.09507 &amp;quot;Deep convolutional neural networks for predominant instrument recognition in polyphonic music&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Hardware Implementations - FPGA, GPU, etc ==&lt;br /&gt;
&lt;br /&gt;
https://www.cse.iitk.ac.in/users/isaha/Publications/Journals/NC10.pdf --&lt;br /&gt;
&amp;quot;Artificial neural networks in hardware: A survey of two decades of progress&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.133.9185&amp;amp;rep=rep1&amp;amp;type=pdf&lt;br /&gt;
&amp;quot;A Self-Repairing Multiplexer-Based FPGA Inspired by Biological Processes&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.genetic-programming.com/jkpdf/fpga1998.pdf -- &amp;quot;Evolving Computer Programs using Rapidly Reconfigurable Field-Programmable Gate Arrays and Genetic Programming&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.114.2588&amp;amp;rep=rep1&amp;amp;type=pdf -- &amp;quot;Flexible Implementation of Genetic Algorithms on FPGAs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.users.muohio.edu/jamiespa/html_papers/gem_10.pdf -- &amp;quot;Revisiting Genetic Algorithms for the FPGA Placement Problem&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.09296v1 -- &amp;quot;Comprehensive Evaluation of OpenCL-based Convolutional Neural Network Accelerators in Xilinx and Altera FPGAs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.409.7533&amp;amp;rep=rep1&amp;amp;type=pdf -- &amp;quot;FPGA-TARGETED NEURAL ARCHITECTURE FOR EMBEDDED ALERTNESS DETECTION&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1611.02450v1 -- &amp;quot;PipeCNN: An OpenCL-Based FPGA Accelerator for Large-Scale Convolution Neuron Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1605.06402v1 -- &amp;quot;Ristretto: Hardware-Oriented Approximation of Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://homes.cs.washington.edu/~luisceze/publications/snnap-hpca-2015.pdf -- &amp;quot;SNNAP: Approximate Computing on Programmable SoCs via Neural Acceleration&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1701.00485v2 -- &amp;quot;Two-Bit Networks for Deep Learning on Resource-Constrained Embedded Devices&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== VLSI ==&lt;br /&gt;
&lt;br /&gt;
http://ncs.ethz.ch/pubs/pdf/Indiveri_etal06.pdf -- &amp;quot;A VLSI array of low-power spiking neurons and bistable synapses with spike–timing dependent plasticity&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Pruning ==&lt;br /&gt;
&lt;br /&gt;
http://papers.nips.cc/paper/5784-learning-both-weights-and-connections-for-efficient-neural-network.pdf --&lt;br /&gt;
&amp;quot;Learning both Weights and Connections for Efficient Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1701.04465 -- &amp;quot;The Incredible Shrinking Neural Network: New Perspectives on Learning Representations Through The Lens of Pruning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1512.08571 -- &amp;quot;Structured Pruning of Deep Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1611.01427 -- &amp;quot;Sparsely-Connected Neural Networks: Towards Efficient VLSI Implementation of Deep Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Efficient Neural Networks via Compression, Quantization, Model Reduction, etc ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1504.04788 -- &amp;quot;Compressing Neural Networks with the Hashing Trick&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1509.08745 -- &amp;quot;Compression of Deep Neural Networks on the Fly&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1502.03436 -- &amp;quot;An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1510.00149 -- &amp;quot;Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.00891 -- &amp;quot;Parameter Compression of Recurrent Neural Networks and Degradation of Short-term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.07061 -- &amp;quot;Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1607.05418 -- &amp;quot;Runtime Configurable Deep Neural Networks for Energy-Accuracy Trade-off&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1602.08194 -- &amp;quot;Scalable and Sustainable Deep Learning via Randomized Hashing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1508.05463 -- &amp;quot;StochasticNet: Forming Deep Neural Networks via Stochastic Connectivity&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1412.7024 -- &amp;quot;Training Deep Neural Networks with Low Precision Multiplications&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.03940 -- &amp;quot;Understanding the Impact of Precision Quantization on the Accuracy and Energy of Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.00222 -- &amp;quot;Ternary Neural Networks for Resource-Efficient AI Applications&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neural Network Hyperparameter Optimization ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1601.00917 -- &amp;quot;DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks&amp;quot; &lt;br /&gt;
&lt;br /&gt;
== Neural Network based EEG Analysis ==&lt;br /&gt;
end&lt;br /&gt;
http://www.seas.upenn.edu/~wulsin/docs/wulsin2010.pdf -- &amp;quot;Semi-Supervised Anomaly Detection for EEG Waveforms Using Deep Belief Nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://inter-eng.upm.ro/2012/files/proceedings/papers/paper72.pdf --&lt;br /&gt;
&amp;quot;Neural Network Parallelization on FPGA Platform for EEG Signal Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Seizure Detection ==&lt;br /&gt;
&lt;br /&gt;
also see https://noisebridge.net/wiki/Kaggle for a (September 2016) current project!&lt;br /&gt;
&lt;br /&gt;
and https://github.com/kevinjos/kaggle-aes-seizure-prediction (some earlier exploration, November 2014)&lt;br /&gt;
&lt;br /&gt;
(broken link, sorry) http://www.sersc.org/journals/ijsip/vol7_no5/26.pdf --&lt;br /&gt;
&amp;quot;A Neural Network Model for Predicting Epileptic Seizures based on Fourier-Bessel Functions&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2230664/pdf/nihms-34528.pdf --&lt;br /&gt;
&amp;quot;A Stochastic Framework for Evaluating Seizure Prediction Algorithms Using Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
(another broken link) http://cs.uni-muenster.de/Professoren/Lippe/diplomarbeiten/html/eisenbach/Untersuchte%20Artikel/PPHD+00.pdf --&lt;br /&gt;
&amp;quot;Recurrent neural network based preenddiction of epileptic seizures in intra- and extracranial EEG&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Visible Light Sensor Network ==&lt;br /&gt;
&lt;br /&gt;
http://infoteh.etf.unssa.rs.ba/zbornik/2016/radovi/KST-1/KST-1-15.pdf --&lt;br /&gt;
&amp;quot;Analysis of Visible Light Communication System for Implementation in Sensor Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neurophysiology ==&lt;br /&gt;
&lt;br /&gt;
http://www.buzsakilab.com/content/PDFs/BuzsakiKoch2012.pdf -- &amp;quot;The origin of extracellular fields and&lt;br /&gt;
currents — EEG, ECoG, LFP and spikes&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Signal Processing ==&lt;br /&gt;
&lt;br /&gt;
http://provideyourown.com/2011/analogwrite-convert-pwm-to-voltage/ -- &amp;quot;Arduino’s AnalogWrite – Converting PWM to a Voltage&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://sim.okawa-denshi.jp/en/PWMtool.php -- &amp;quot;RC Low-pass Filter Design for PWM (Transient Analysis Calculator)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Hyperdimensional Computing ==&lt;br /&gt;
&lt;br /&gt;
http://redwood.berkeley.edu/pkanerva/papers/kanerva09-hyperdimensional.pdf --&lt;br /&gt;
&amp;quot;Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1602.03032.pdf --&lt;br /&gt;
&amp;quot;Associative Long Short-Term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Bird Flocks and Maximum Entropy ==&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1107.0604v1 --&lt;br /&gt;
&amp;quot;Statistical Mechanics and Flocks of Birds&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1307.5563v1 --&lt;br /&gt;
&amp;quot;Social interactions dominate speed control in driving natural flocks toward criticality&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ocw.mit.edu/courses/electrical-engineering-and-computer-science/6-050j-information-and-entropy-spring-2008/syllabus/MIT6_050JS08_penfield.pdf --&lt;br /&gt;
&amp;quot;Information and Entropy (Course Notes)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Whale Songs ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1307.0589.pdf -- &amp;quot;The Orchive : Data mining a massive bioacoustic archive&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.researchgate.net/profile/Herbert_Roitblat/publication/13429327_The_neural_network_classification_of_false_killer_whale_%28Pseudorca_crassidens%29_vocalizations/links/540d2ff60cf2df04e75478cd.pdf -- &amp;quot;The neural network classification of false killer whale (Pseudorca crassidens) vocalizations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://users.iit.demokritos.gr/~paliourg/papers/PhD.pdf -- &amp;quot;REFINEMENT OF TEMPORAL CONSTRAINTS IN AN EVENT RECOGNITION SYSTEM USING SMALL DATASETS&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.nersc.no/sites/www.nersc.no/files/master_thesis_sebastian_menze.pdf -- &amp;quot;Estimating fin whale distribution from ambient noise spectra using Bayesian inversion&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://sis.univ-tln.fr/~glotin/IJCNN2015_IHMMbioac_BartChamGlot.pdf -- &amp;quot;Hierarchical Dirichlet Process Hidden Markov Model for Unsupervised Bioacoustic Analysis&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.inf.ed.ac.uk/publications/thesis/online/IM030057.pdf -- &amp;quot;Hidden Markov Model Clustering of Acoustic Data&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://danielnouri.org/notes/2014/01/10/using-deep-learning-to-listen-for-whales/ -- Using deep learning to listen for whales&lt;br /&gt;
&lt;br /&gt;
== Computational Cognitive Neuroscience ==&lt;br /&gt;
&lt;br /&gt;
http://www.pnas.org/content/110/41/16390.full -- &amp;quot;Indirection and symbol-like processing in the prefrontal cortex and basal ganglia&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ruccs.rutgers.edu/images/personal-zenon-pylyshyn/docs/jaf.pdf -- &amp;quot;Connectionism and Cognitive Architecture: A Critical Analysis&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2728678/pdf/nihms131814.pdf -- Neves et al 2008 &amp;quot;Cell Shape and Negative Links in Regulatory Motifs Together Control Spatial Information Flow in Signaling Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://psych.colorado.edu/~oreilly/papers/AisaMingusOReilly08.pdf -- &amp;quot;The Emergent Neural Modeling System&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://grey.colorado.edu/emergent/index.php/Comparison_of_Neural_Network_Simulators&lt;br /&gt;
&lt;br /&gt;
== Text Generation ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~ilya/pubs/2011/LANG-RNN.pdf -- &amp;quot;Generating Text with Recurrent Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Games ==&lt;br /&gt;
&lt;br /&gt;
http://setgame.com/sites/default/files/teacherscorner/COGNITIVE%20MODELING%20WITH%20SET.pdf -- &amp;quot;How to Construct a Believable Opponent using Cognitive Modeling in the Game of Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www-personal.umich.edu/~charchan/SET.pdf -- &amp;quot;SETs and Anti-SETs: The Math Behind the Game of SET&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://personal.plattsburgh.edu/quenelgt/talks/set.pdf -- &amp;quot;Introduction to Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://web.engr.illinois.edu/~pbg/papers/set.pdf -- &amp;quot;On the Complexity of the Game of Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.warwick.ac.uk/staff/D.Maclagan/papers/set.pdf -- &amp;quot;The Card Game Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.math.ucdavis.edu/~anne/FQ2011/set_game.pdf -- &amp;quot;The Game Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Large Scale Brain Simulation ==&lt;br /&gt;
&lt;br /&gt;
http://www.nowere.net/b/arch/96550/src/1378907656268.pdf -- &amp;quot;A world survey of artificial brain projects, Part I: Large-scale brain simulations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Music ==&lt;br /&gt;
&lt;br /&gt;
http://cmr.soc.plymouth.ac.uk/publications/bci-wkshop.pdf -- &amp;quot;ON GENERATING EEG FOR CONTROLLING MUSICAL SYSTEMS&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
https://github.com/nbdt/gotrain (our ANN code)&lt;br /&gt;
&lt;br /&gt;
https://github.com/nbdt/openbci-golang-server (EEG data acquisition and visualization software for OpenBCI)&lt;br /&gt;
&lt;br /&gt;
== Hidden Markov Models ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.sjsu.edu/~stamp/RUA/HMM.pdf&lt;br /&gt;
-- &amp;quot;A Revealing Introduction to Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.jelmerborst.nl/pubs/Borst2013b.pdf&lt;br /&gt;
-- &amp;quot;Discovering Processing Stages by combining EEG with Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2230664/pdf/nihms-34528.pdf&lt;br /&gt;
-- &amp;quot;A Stochastic Framework for Evaluating Seizure Prediction Algorithms Using Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.schalklab.org/sites/default/files/misc/Coupled%20hidden%20markov%20model%20for%20electrocorticographic%20signal%20classification.pdf&lt;br /&gt;
-- &amp;quot;Coupled Hidden Markov Model for Electrocorticographic Signal Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Long Short Term Memory ==&lt;br /&gt;
&lt;br /&gt;
http://deeplearning.cs.cmu.edu/pdfs/Hochreiter97_lstm.pdf&lt;br /&gt;
-- &amp;quot;Long Short-Term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
ftp://ftp.idsia.ch/pub/juergen/2002_icannMusic.pdf&lt;br /&gt;
-- &amp;quot;Learning The Long-Term Structure of the Blues&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.overcomplete.net/papers/nn2012.pdf&lt;br /&gt;
-- &amp;quot;A generalized LSTM-like training algorithm for second-order recurrent neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.iro.umontreal.ca/~eckdoug/papers/2003_nn.pdf&lt;br /&gt;
-- &amp;quot;Kalman filters improve LSTM network performance in problems unsolvable by traditional recurrent nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://googleresearch.blogspot.ch/2015/09/google-voice-search-faster-and-more.html&lt;br /&gt;
-- &amp;quot;Long Short-Term Memory dramatically improves Google Voice etc&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1511.05552v4.pdf --&lt;br /&gt;
&amp;quot;Recurrent Neural Networks Hardware Implementation on FPGA&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://vast.cs.ucla.edu/sites/default/files/publications/ASP-DAC2017-1352-11.pdf --&lt;br /&gt;
&amp;quot;FPGA-based Accelerator for Long Short-Term Memory Recurrent Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Question Answering ==&lt;br /&gt;
&lt;br /&gt;
http://www.overcomplete.net/papers/bica2012.pdf&lt;br /&gt;
-- &amp;quot;Neural Architectures for Learning to Answer Questions&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://cs.umd.edu/~miyyer/pubs/2014_qb_rnn.pdf&lt;br /&gt;
-- &amp;quot;A Neural Network for Factoid Question Answering over Paragraphs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1502.05698.pdf&lt;br /&gt;
-- &amp;quot;Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.peterschueller.com/pub/2014/2014_winograd_schemas_relevance_knowledge_graphs.pdf&lt;br /&gt;
-- &amp;quot;Tackling Winograd Schemas by Formalizing Relevance Theory in Knowledge Graphs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ijcai.org/papers15/Papers/IJCAI15-190.pdf&lt;br /&gt;
-- &amp;quot;Towards Addressing the Winograd Schema Challenge — Building and Using a Semantic Parser and a Knowledge Hunting Module&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1506.05869v2.pdf&lt;br /&gt;
-- &amp;quot;A Neural Conversational Model&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1508.05508v1.pdf&lt;br /&gt;
-- &amp;quot;Towards Neural Network-based Reasoning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.visualqa.org/vqa_iccv2015.pdf&lt;br /&gt;
-- &amp;quot;VQA: Visual Question Answering&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Propagators ==&lt;br /&gt;
Cells must support three operations:&lt;br /&gt;
*add some content&lt;br /&gt;
*collect the content currently accumulated&lt;br /&gt;
*register a propagator to be notified when the accumulated content changes&lt;br /&gt;
*When new content is added to a cell, the cell must merge the addition with the content already present. When a propagator asks for the content of a cell, the cell must deliver a complete summary of the information that has been added to it.&lt;br /&gt;
*The merging of content must be commutative, associative, and idempotent. The behavior of propagators must be monotonic with respect to the lattice induced by the merge operation.&lt;br /&gt;
*http://web.mit.edu/~axch/www/art.pdf *http://groups.csail.mit.edu/mac/users/gjs/propagators/&lt;br /&gt;
*http://dustycloud.org/blog/sussman-on-ai/&lt;br /&gt;
&lt;br /&gt;
== Boosting ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.princeton.edu/courses/archive/spr08/cos424/readings/Schapire2003.pdf&lt;br /&gt;
-- &amp;quot;The Boosting Approach to Machine Learning An Overview&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.19.9955&amp;amp;rep=rep1&amp;amp;type=pdf&lt;br /&gt;
-- &amp;quot;Ensembling Neural Networks: Many Could Be Better Than All&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cs.columbia.edu/~rocco/Public/mlj9.pdf&lt;br /&gt;
-- &amp;quot;Random Classification Noise Defeats All Convex Potential Boosters&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Support Vector Machines ==&lt;br /&gt;
&lt;br /&gt;
http://research.microsoft.com/en-us/um/people/cburges/papers/svmtutorial.pdf&lt;br /&gt;
-- &amp;quot;A Tutorial on Support Vector Machines for Pattern Recognition&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Wire Length / Small World Networks ==&lt;br /&gt;
&lt;br /&gt;
http://koulakovlab.cshl.edu/publications/koulakov_chklovskii_2000.pdf&lt;br /&gt;
-- &amp;quot;A wire length minimization approach to ocular dominance patterns in mammalian visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://papers.nips.cc/paper/1910-foundations-for-a-circuit-complexity-theory-of-sensory-processing.pdf&lt;br /&gt;
-- &amp;quot;Foundations for a Circuit Complexity Theory of Sensory Processing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.nada.kth.se/~cjo/documents/small_world.pdf&lt;br /&gt;
-- &amp;quot;Small-World Connectivity and Attractor Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.doc.ic.ac.uk/~mpsha/ShanahanPhysRevEPreprint.pdf&lt;br /&gt;
-- &amp;quot;The Dynamical Complexity of Small-World Networks of Spiking Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dam.brown.edu/people/elie/papers/small_world.pdf&lt;br /&gt;
-- &amp;quot;Autoassociative Memory Retrieval and Spontaneous Activity Bumps in Small-World Networks of Integrate-and-Fire Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.332.4501&amp;amp;rep=rep1&amp;amp;type=pdf&lt;br /&gt;
-- &amp;quot;Transition from Random to Small-World Neural Networks by STDP Learning Rule&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://tamar.tau.ac.il/~eshel/papers/J_Neural_Engineering.pdf&lt;br /&gt;
-- &amp;quot;Compact self-wiring in cultured neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Backpropagation ==&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/neuron.pdf -- &amp;quot;Neural Networks - A Systematic Introduction&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www4.rgu.ac.uk/files/chapter3%20-%20bp.pdf (works through simple example)&lt;br /&gt;
&lt;br /&gt;
http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example/&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/chapter/K7.pdf&lt;br /&gt;
&lt;br /&gt;
also see http://page.mi.fu-berlin.de/rojas/neural/neuron.pdf (entire book &amp;quot;Neural Networks - a Systemic Introduction&amp;quot; by Raul Rojas)&lt;br /&gt;
&lt;br /&gt;
http://work.caltech.edu/lectures.html Hoeffding&#039;s inequality, VC Dimension and Back Propagation ANN&lt;br /&gt;
&lt;br /&gt;
http://www.cs.stir.ac.uk/research/publications/techreps/pdf/TR148.pdf (&amp;quot;Learning XOR: exploring the space of a classic problem&amp;quot;)&lt;br /&gt;
&lt;br /&gt;
http://www.sysc.pdx.edu/classes/Werbos-Backpropagation%20through%20time.pdf&lt;br /&gt;
-- &amp;quot;Backpropagation Through Time: What it Does and How to Do It&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Computer Vision ==&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/CVPR2015.py -- some interesting papers here&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Nguyen_Deep_Neural_Networks_2015_CVPR_paper.pdf -- &amp;quot;Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Visual Perception (Biological Systems) ==&lt;br /&gt;
&lt;br /&gt;
http://cbcl.mit.edu/publications/ps/Serre_etal_PBR07.pdf&lt;br /&gt;
-- &amp;quot;A quantitative theory of immediate visual recognition&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dam.brown.edu/ptg/REPORTS/Invariance.pdf&lt;br /&gt;
-- &amp;quot;Invariance and Selectivity in the Ventral Visual Pathway&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cnbc.cmu.edu/~tai/papers/lee_mumford_josa.pdf&lt;br /&gt;
-- &amp;quot;Hierarchical Bayesian inference in the visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neural Synchrony ==&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1312.6115.pdf&lt;br /&gt;
-- &amp;quot;Neuronal Synchrony in Complex-Valued Deep Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Spiking Neural Networks ==&lt;br /&gt;
&lt;br /&gt;
http://ncs.ethz.ch/projects/evospike/publications/ICONIP2011%20Springer%20LNCS%20Nutta.pdf -- &amp;quot;EEG Classification with BSA Spike Encoding Algorithm and Evolving Probabilistic Spiking Neural Network&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://dai.fmph.uniba.sk/~benus/confers/CNGM_ICANN05.pdf -- &amp;quot;Computational Neurogenetic Modeling: Integration of Spiking Neural Networks, Gene Networks, and Signal Processing Techniques&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.97.3902 -- &amp;quot;Pattern Recognition in a Bucket&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.igi.tugraz.at/maass/psfiles/221.pdf -- &amp;quot;Noise as a Resource for Computation and Learning in Spiking Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://web.stanford.edu/group/brainsinsilicon/goals.html -- Neurogrid&lt;br /&gt;
&lt;br /&gt;
http://apt.cs.manchester.ac.uk/projects/SpiNNaker/ -- SpiNNaker&lt;br /&gt;
&lt;br /&gt;
http://personalpages.manchester.ac.uk/staff/javier.navaridas/pubs/2012.jpdc.pdf -- more on SpiNNaker&lt;br /&gt;
&lt;br /&gt;
==Hierarchical Temporal Memory==&lt;br /&gt;
&lt;br /&gt;
https://github.com/numenta/nupic/wiki/Hierarchical-Temporal-Memory-Theory&lt;br /&gt;
&lt;br /&gt;
http://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1000532 -- &amp;quot;Towards a Mathematical Theory of Cortical Micro-circuits&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Distributed Neural Networks==&lt;br /&gt;
&lt;br /&gt;
https://www.paypal-engineering.com/2015/01/12/deep-learning-on-hadoop-2-0-2/ -- paypal engineering blog, paper describes implementation of neural network as described by [http://www.maths.bris.ac.uk/~maxvd/reduce_dim.pdf Hinton 2006] on Hadoop&lt;br /&gt;
&lt;br /&gt;
http://www.sciencemag.org/content/suppl/2006/08/04/313.5786.504.DC1/Hinton.SOM.pdf -- supplementary material for [http://www.maths.bris.ac.uk/~maxvd/reduce_dim.pdf Hinton 2006]&lt;br /&gt;
&lt;br /&gt;
http://www.cs.otago.ac.nz/staffpriv/hzy/papers/pdcs03.pdf -- &amp;quot;Parallelization of a Backpropagation Neural Network on a Cluster Computer&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1404.5997v2.pdf -- &amp;quot;One weird trick for parallelizing convolutional neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ics.uci.edu/~jmoorkan/pub/gpusnn-ijcnn.pdf -- &amp;quot;Efficient Simulation of Large-Scale Spiking Neural Networks Using CUDA Graphics Processors&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Mixture of Experts==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~fritz/absps/jjnh91.pdf -- &amp;quot;Adaptive Mixtures of Local Experts&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Hopfield nets and RBMs==&lt;br /&gt;
&lt;br /&gt;
http://www.iro.umontreal.ca/~bengioy/papers/ftml_book.pdf -- Yoshua Bengio, 2009, Learning Deep Architectures for AI&lt;br /&gt;
&lt;br /&gt;
http://deeplearning.cs.cmu.edu/ -- Syllabus for cs course on deep learning, possible source of literature for the library &lt;br /&gt;
&lt;br /&gt;
https://class.coursera.org/neuralnets-2012-001/lecture -- Lecture 11 - 15 to aid in understanding of Hinton 2006 paper above&lt;br /&gt;
&lt;br /&gt;
http://www.pnas.org/content/79/8/2554.full.pdf -- &amp;quot;Neural networks and physical systems with emergent collective computational abilities&amp;quot; (Hopfield 1982)&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/chapter/K13.pdf -- &amp;quot;The Hopfield Model&amp;quot; (Rojas 1996)&lt;br /&gt;
&lt;br /&gt;
http://www.seas.upenn.edu/~wulsin/docs/wulsin2010.pdf -- &amp;quot;Semi-Supervised Anomaly Detection for EEG Waveforms Using Deep Belief Nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cse.buffalo.edu/DBGROUP/bioinformatics/papers/cameraready_xwjiabibe.pdf -- &amp;quot;A Novel Semi-supervised Deep Learning Framework&lt;br /&gt;
for Affective State Recognition on EEG Signals&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~hinton/absps/guideTR.pdf -- &amp;quot;A Practical Guide to Training Restricted Boltzmann Machines&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1503.07793v2.pdf&lt;br /&gt;
-- &amp;quot;Gibbs Sampling with Low-Power Spiking Digital Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1311.0190v1 -- &amp;quot;On the typical properties of inverse problems in statistical mechanics&amp;quot; Iacopo Mastromatteo 2013&lt;br /&gt;
&lt;br /&gt;
http://jmlr.org/proceedings/papers/v5/salakhutdinov09a/salakhutdinov09a.pdf -- &amp;quot;Deep Boltzmann Machines&amp;quot; Salakhutdinov &amp;amp; Hinton 2009&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~hinton/absps/tr00-004.pdf -- &amp;quot;Training Products of Experts by Minimizing Contrastive Divergence&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.eecg.toronto.edu/~pc/research/publications/ly.fpga2009.submitted.pdf -- &amp;quot;A High-Performance FPGA Architecture for Restricted&lt;br /&gt;
Boltzmann Machines&amp;quot; Ly &amp;amp; Chow 2009&lt;br /&gt;
&lt;br /&gt;
https://pdfs.semanticscholar.org/85fa/f7c3c05388e2bcd097a416606bdd88fc0c7c.pdf -- &amp;quot;A MULTI-FPGA ARCHITECTURE FOR STOCHASTIC RESTRICTED BOLTZMANN MACHINES&amp;quot; Ly &amp;amp; Chow 2009&lt;br /&gt;
&lt;br /&gt;
== Neuromorphic Stuff ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1508.01008.pdf -- &amp;quot;INsight: A Neuromorphic Computing System for Evaluation of Large Neural Networks&amp;quot; Chung, Shin &amp;amp; Kang 2015&lt;br /&gt;
&lt;br /&gt;
== Markov Chain Monte Carlo ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.princeton.edu/courses/archive/spr06/cos598C/papers/AndrieuFreitasDoucetJordan2003.pdf&lt;br /&gt;
-- &amp;quot;An Introduction to MCMC for Machine Learning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://jmlr.org/proceedings/papers/v37/salimans15.pdf&lt;br /&gt;
-- &amp;quot;Markov Chain Monte Carlo and Variational Inference: Bridging the Gap&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.umiacs.umd.edu/~resnik/pubs/LAMP-TR-153.pdf -- &amp;quot;Gibbs Sampling for the Uninitiated&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Entrainment==&lt;br /&gt;
&lt;br /&gt;
http://www.scirp.org/Journal/PaperDownload.aspx?paperID=33251 -- &amp;quot;Alpha Rhythms Response to 10 Hz Flicker Is Wavelength Dependent&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.brainmachine.co.uk/wp-content/uploads/Herrmann_Flicker.pdf -- &amp;quot;EEG responses to 1–100 Hz flicker: resonance phenomena in visual cortex and their potential correlation to cognitive phenomena&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.jneurosci.org/content/23/37/11621.full.pdf -- &amp;quot;Human Cerebral Activation during Steady-State Visual-Evoked Responses&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dauwels.com/Papers/CogDyn%202009.pdf -- &amp;quot;On the synchrony of steady state visual evoked potentials and oscillatory burst events&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.tu-ilmenau.de/fileadmin/public/lorentz-force/publications/peer/2012/haueisen2012/Halbleib_JCN_2012_Topographic_analysis_photic_driving.pdf -- &amp;quot;Topographic Analysis of Engagement and Disengagement of Neural Oscillators in Photic Driving: A Combined Electroencephalogram/Magnetoencephalogram Study&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Mining Scientific Literature==&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC153503/pdf/1471-2105-4-11.pdf -- &amp;quot;PreBIND and Textomy – mining the biomedical literature for protein-protein interactions using a support vector machine&amp;quot; Donaldson 2003 BMC Bioinformatics&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4674139/pdf/pcbi.1004630.pdf -- &amp;quot;Text Mining for Protein Docking&amp;quot; Badal 2015 PLoS Comput Biol.&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4691339/pdf/bav116.pdf -- &amp;quot;Biocuration with insufficient resources and fixed timelines&amp;quot; Rodriguez-Esteban 2015 Database: The Journal of Biological Databases and Curation&lt;br /&gt;
&lt;br /&gt;
==(not necessarilly very) Current Discussion==&lt;br /&gt;
&lt;br /&gt;
re Tononi&#039;s &amp;quot;Integrated Information Theory&amp;quot; http://www.scottaaronson.com/blog/?p=1799&lt;br /&gt;
&lt;br /&gt;
(19 February 2014) starting to think about possibility for experiments (loosely) related to [https://en.wikipedia.org/wiki/Visual_evoked_potential Visual Evoked Potential] research again - for instance:&lt;br /&gt;
&lt;br /&gt;
http://www.biosemi.com/publications/pdf/Bierman.pdf -- Wave function collapse&lt;br /&gt;
&lt;br /&gt;
[[File:Schak99InstantaneousCoherenceStroopTask.pdf]] -- &amp;quot;Instantaneous EEG Coherence Analysis During the Stroop Task&amp;quot; -- Schack et al 1999&lt;br /&gt;
&lt;br /&gt;
[[File:CoherentEEGAmbiguousFigureBinding.pdf]] -- &amp;quot;Coherent EEG Indicators of Cognitive Binding During Ambiguous Figure Tasks&amp;quot; -- Klemm, Li, and Hernandez 2000 &lt;br /&gt;
&lt;br /&gt;
Note these two papers flog coherence measures - not trying to focus so much on that analysis right now, more interested in general understanding of what these experiments are about with possible goal of designing simpler experiments &amp;amp; analysis of similar perceptual/cognitive phenomena.&lt;br /&gt;
&lt;br /&gt;
Here is an article that looks more directly at visual evoked potential measures:&lt;br /&gt;
&lt;br /&gt;
[[File:ERP_Stereoscopic.pdf]] -- &amp;quot;Evoked Responses to Distinct and Nebulous Stereoscopic Stimuli&amp;quot; -- Dunlop et al 1983&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
(11 September 2013) more on analysis methods:&lt;br /&gt;
&lt;br /&gt;
http://slesinsky.org/brian/misc/eulers_identity.html&lt;br /&gt;
&lt;br /&gt;
http://www.dspguide.com/ch8/1.htm&lt;br /&gt;
&lt;br /&gt;
[[File:Fftw3.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ParametricEEGAnalysis.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ICATutorial.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ICAFrequencyDomainEEG.pdf]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
(21 August 2013) - readings relating statistical (etc math / signal processing / pattern recognition / machine learning) methods for EEG data interpretation.  A lot of stuff, a bit of nonsense ... and ... statistics!&lt;br /&gt;
&lt;br /&gt;
Would be good to identify any papers suitable for more in-depth study.  Currently have a wide field to graze for selections:&lt;br /&gt;
&lt;br /&gt;
[[File:DWTandFFTforEEG.pdf]] &amp;quot;EEG Classifier using Fourier Transform and Wavelet Transform&amp;quot; -- Maan Shaker, 2007&lt;br /&gt;
&lt;br /&gt;
[[File:Schak99InstantaneousCoherenceStroopTask.pdf]] -- &amp;quot;Instantaneous EEG Coherence Analysis During the Stroop Task&amp;quot; -- Schack et al 1999&lt;br /&gt;
&lt;br /&gt;
[[File:KulaichevCoherence.pdf]] -- &amp;quot;The Informativeness of Coherence Analysis in EEG Studies&amp;quot; -- A. P. Kulaichev 2009 &#039;&#039;note: interesting critical perspective re limitations, discussion of alternative analytics&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
[[File:ContinuousAndDiscreteWaveletTransforms.pdf]] -- review of (pre-1990) wavelet literature -- Christopher Heil and David Walnut, 1989&lt;br /&gt;
&lt;br /&gt;
[[File:EEGGammaMeditation.pdf]] -- &amp;quot;Brain sources of EEG gamma frequency during volitionally meditation-induced, altered states of consciousness, and experience of the self&amp;quot; -- Dietrich Lehman et al 2001&lt;br /&gt;
&lt;br /&gt;
http://neuro.hut.fi/~pavan/home/Hyvarinen2010_FourierICA_Neuroimage.pdf - &amp;quot;Independent component analysis of short-time Fourier transforms for spontaneous EEG/MEG analysis&amp;quot; -- Aapo Hyvarinen, Pavan Ramkumar, Lauri Parkkonen, Riitta Hari - paper published in Neuroimage vol 49 (2010)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://www.nickgillian.com/software/grt OpenSource Machine Learning Algs from NG @MIT]&lt;br /&gt;
&amp;lt;br&amp;gt;[https://www.usenix.org/system/files/conference/usenixsecurity12/sec12-final56.pdf Consumer grade EEG used to see &amp;quot;P300&amp;quot; reponse] and for thoes with a short attention span [http://www.extremetech.com/extreme/134682-hackers-backdoor-the-human-brain-successfully-extract-sensitive-data tldr]&lt;br /&gt;
&amp;lt;br&amp;gt;(discussed at meetup Wednesday 31 July 2013)&lt;br /&gt;
&amp;lt;br&amp;gt;&amp;quot;Coherent EEG Indicators of Cognitive Binding During Ambiguous Figure Tasks&amp;quot; Klemm, Li, and Hernandez 2000 &lt;br /&gt;
&amp;lt;br&amp;gt;[[File:CoherentEEGAmbiguousFigureBinding.pdf]]&lt;br /&gt;
&amp;lt;br&amp;gt;&amp;quot;We tested the hypothesis that perception of an alternative image in ambiguous figures would be manifest as high-frequency (gamma) components that become synchronized over multiple scalp sites as a &amp;quot;cognitive binding&amp;quot; process occurs.&amp;quot;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://dreamsessions.net art, dream, and eeg]&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://www.believermag.com/issues/200710/?read=article_aviv mind v brain, hobson v solms]&lt;br /&gt;
&amp;lt;br&amp;gt;http://en.wikipedia.org/wiki/Activation-synthesis_hypothesis&lt;br /&gt;
&amp;lt;br&amp;gt;[[File:HobsomREMDreamProtoconsciousness.pdf|Hobson09ProtosconsciousnessREMDream]]&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Hobson and McCarley originally proposed in the 1970s that the differences in the waking-NREM-REM sleep cycle was the result of interactions between aminergic REM-off cells and cholinergic REM-on cells.[4] This was perceived as the activation-synthesis model, stating that brain activation during REM sleep results in synthesis of dream creation.[1][1] Hobson&#039;s five cardinal characteristics include: intense emotions, illogical content, apparent sensory impressions, uncritical acceptance of dream events, and difficulty in being remembered.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
Berkeley Labs&lt;br /&gt;
&lt;br /&gt;
[http://gallantlab.org/index.html Gallant Group]&lt;br /&gt;
&amp;lt;br&amp;gt;[http://walkerlab.berkeley.edu/ Walker Group]&lt;br /&gt;
&amp;lt;br&amp;gt;[http://socrates.berkeley.edu/~plab/ Palmer Group]&lt;br /&gt;
&lt;br /&gt;
==Sleep Research==&lt;br /&gt;
&lt;br /&gt;
[http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2335403/ Comment on the AASM Manual for the Scoring of Sleep and Associated Events]&lt;br /&gt;
&lt;br /&gt;
==random tangents==&lt;br /&gt;
(following previous discussion) - we might select a few to study in more depth&lt;br /&gt;
(... or not!  Plenty more to explore - suggestions (random or otherwise) are welcome.&lt;br /&gt;
http://www.meltingasphalt.com/neurons-gone-wild/ --&lt;br /&gt;
Neurons Gone Wild - Levels of agency in the brain. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;stereoscopic perception:&#039;&#039;&#039;&lt;br /&gt;
*[[File:ERP_Stereoscopic.pdf]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
some (maybe) interesting background on Information Theory (cool title...)&lt;br /&gt;
 Claude Shannon: &amp;quot;Communication in the Presence of Noise&amp;quot;&lt;br /&gt;
 [[File:Shannon_noise.pdf]]&lt;br /&gt;
 &amp;quot;We will call a system that transmits without errors at the rate &#039;&#039;C&#039;&#039; an ideal system.&lt;br /&gt;
  Such a system cannot be achieved with any finite encoding process&lt;br /&gt;
  but can be approximated as closely as desired.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
wikipedia etc quick reads:&lt;br /&gt;
 https://en.wikipedia.org/wiki/Eeg&lt;br /&gt;
 https://en.wikipedia.org/wiki/Neural_synchronization&lt;br /&gt;
 https://en.wikipedia.org/wiki/Event-related_potentials&lt;br /&gt;
 http://www.scholarpedia.org/article/Spike-and-wave_oscillations&lt;br /&gt;
 http://www.scholarpedia.org/article/Thalamocortical_oscillations&lt;br /&gt;
&lt;br /&gt;
==Previously==&lt;br /&gt;
&lt;br /&gt;
[http://www.psychiclab.net/ Masahiro&#039;s EEG Device/IBVA Software]&lt;br /&gt;
&lt;br /&gt;
[http://www.instructables.com/id/open-brain-wave-interface-hardware-1/ and ... open source hardware design and kits on instructables.com]&lt;br /&gt;
&lt;br /&gt;
[http://brainstorms.puzzlebox.info/ Puzzlebox - Opensource BCI Developers]&lt;br /&gt;
&lt;br /&gt;
Morgan from GazzLab @ MissionBay/UCSF&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
https://github.com/gazzlab&lt;br /&gt;
&lt;br /&gt;
Let&#039;s ease into a lightweight &amp;quot;journal club&amp;quot; discussion with this technical report from NeuroSky.&lt;br /&gt;
&lt;br /&gt;
Name: A user-friendly SSVEP-based brain-computer interface using a time-domain classifier, Luo A and Sullivan TJ 2010&lt;br /&gt;
&lt;br /&gt;
URL: [[File:NeuroSkyVEP.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
Please add your comments &amp;amp; questions here.&lt;br /&gt;
&lt;br /&gt;
==Background Reading==&lt;br /&gt;
&lt;br /&gt;
http://nanosouffle.net/ (view into Arxiv.org)&lt;br /&gt;
&lt;br /&gt;
Name: Hunting for Meaning after Midnight, Miller 2007&lt;br /&gt;
&lt;br /&gt;
URL: &amp;lt;http://www.noisebridge.net/pipermail/neuro/attachments/20130501/4b992eb6/attachment-0002.pdf&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Name: Broken mirrors, Ram, VS, &amp;amp; Oberman, LM, 2006, Nov&lt;br /&gt;
&lt;br /&gt;
URL: &amp;lt;http://www.noisebridge.net/pipermail/neuro/attachments/20130501/4b992eb6/attachment-0003.pdf&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Ramachandran Critique&lt;br /&gt;
&lt;br /&gt;
http://blogs.scientificamerican.com/guest-blog/2012/11/06/whats-so-special-about-mirror-neurons/&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2773693/&lt;br /&gt;
&lt;br /&gt;
Sleep/Dream Studies&lt;br /&gt;
&lt;br /&gt;
http://www.cns.atr.jp/dni/en/publications/&lt;br /&gt;
&lt;br /&gt;
==NeuroSky Docs==&lt;br /&gt;
[[File:NeuroSkyDongleProtocol.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
[[File:NeuroSkyCommunicationsProtocol.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
==Android Neutral Network Fuzzy Learning app==&lt;br /&gt;
[https://play.google.com/store/apps/details?id=com.faadooengineers.free_neuralnetworkandfuzzysystems Android Neutral Network Fuzzy Learning app in Play Store]&lt;br /&gt;
&lt;br /&gt;
==Learning about Neural Networks==&lt;br /&gt;
* What type of network? [http://stats.stackexchange.com/questions/114385/what-is-the-difference-between-convolutional-neural-networks-restricted-boltzma RMB (Restricted Boltzmann Machine) vs Autoencoder/MLP vs CNN (Convolutional Neural Networks)]&lt;br /&gt;
* Andrej Karpathy&#039;s [http://cs.stanford.edu/people/karpathy/convnetjs/ Convolutional Neural Network coded in JavaScript (ConvNetJS)]&lt;br /&gt;
* Andrej Karpathy&#039;s [http://karpathy.github.io/2015/10/25/selfie/ What a Deep Neural Network thinks about your #selfie  (background on Convolutional Neural Networks for image recognition and classification)]&lt;br /&gt;
* [https://blog.webkid.io/neural-networks-in-javascript/ Neural Networks in JavaScript w/MNIST]&lt;br /&gt;
* [http://www.antoniodeluca.info/blog/10-08-2016/neural-networks-in-javascript.html Another NN in JS]&lt;br /&gt;
* [http://caza.la/synaptic/ The Synaptic &amp;quot;architecture-free&amp;quot; neural network library in JS]&lt;/div&gt;</summary>
		<author><name>198.27.140.180</name></author>
	</entry>
	<entry>
		<id>https://wiki.extremist.software/index.php?title=DreamTeam/Reading&amp;diff=57968</id>
		<title>DreamTeam/Reading</title>
		<link rel="alternate" type="text/html" href="https://wiki.extremist.software/index.php?title=DreamTeam/Reading&amp;diff=57968"/>
		<updated>2017-04-08T22:44:50Z</updated>

		<summary type="html">&lt;p&gt;198.27.140.180: /* Whale Songs */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;(note wiki contains some useful clues re previous neuro research at Noisebridge ... For example, the [[Analog_EEG_Amp]] page describes some project ideas and work done by others here in 2012) &lt;br /&gt;
&lt;br /&gt;
https://metacademy.org/&lt;br /&gt;
-- machine learning knowledge graph&lt;br /&gt;
&lt;br /&gt;
http://www.thetalkingmachines.com/ -- podcast&lt;br /&gt;
&lt;br /&gt;
https://karpathy.github.io/2015/05/21/rnn-effectiveness/&lt;br /&gt;
&lt;br /&gt;
== Bayesian Inference ==&lt;br /&gt;
&lt;br /&gt;
https://noisebridge.net/images/e/ef/Perception_is_in_the_Details12.pdf --&lt;br /&gt;
&amp;quot;Perception is in the Details: A Predictive Coding Account of the Psychedelic Phenomenon&amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
http://rsif.royalsocietypublishing.org/content/10/86/20130475 --&lt;br /&gt;
&amp;quot;Life as we know it&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://jmlr.csail.mit.edu/proceedings/papers/v31/wang13b.pdf --&lt;br /&gt;
&amp;quot;Collapsed Variational Bayesian Inference for Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.datalab.uci.edu/papers/nips06_cvb.pdf --&lt;br /&gt;
&amp;quot;A Collapsed Variational Bayesian Inference Algorithm for Latent Dirichlet Allocation&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cnbc.cmu.edu/~tai/papers/lee_mumford_josa.pdf --&lt;br /&gt;
&amp;quot;Hierarchical Bayesian inference in the visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.researchgate.net/profile/Til_Bergmann/publication/262423308_Temporal_coding_organized_by_coupled_alpha_and_gamma_oscillations_prioritize_visual_processing/links/0deec537d1bfda474c000000/Temporal-coding-organized-by-coupled-alpha-and-gamma-oscillations-prioritize-visual-processing.pdf --&lt;br /&gt;
&amp;quot;Temporal coding organized by coupled alpha and gamma oscillations prioritize visual processing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cell.com/neuron/pdf/S0896-6273(15)00823-5.pdf --&lt;br /&gt;
&amp;quot;Rhythms for Cognition: Communication through Coherence&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.biorxiv.org/content/biorxiv/early/2014/05/06/004804.full.pdf --&lt;br /&gt;
&amp;quot;Visual Areas Exert Feedforward and Feedback Influences through Distinct Frequency Channels&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Speech Recognition ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.00694v1 -- &amp;quot;ESE: Efficient Speech Recognition Engine with Compressed LSTM on FPGA&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Sound Classification ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1608.04363v2 -- &amp;quot;Deep Convolutional Neural Networks and Data Augmentation for Environmental Sound Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1605.09507 &amp;quot;Deep convolutional neural networks for predominant instrument recognition in polyphonic music&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Hardware Implementations - FPGA, GPU, etc ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.09296v1 -- &amp;quot;Comprehensive Evaluation of OpenCL-based Convolutional Neural Network Accelerators in Xilinx and Altera FPGAs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.409.7533&amp;amp;rep=rep1&amp;amp;type=pdf -- &amp;quot;FPGA-TARGETED NEURAL ARCHITECTURE FOR EMBEDDED ALERTNESS DETECTION&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1611.02450v1 -- &amp;quot;PipeCNN: An OpenCL-Based FPGA Accelerator for Large-Scale Convolution Neuron Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1605.06402v1 -- &amp;quot;Ristretto: Hardware-Oriented Approximation of Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://homes.cs.washington.edu/~luisceze/publications/snnap-hpca-2015.pdf -- &amp;quot;SNNAP: Approximate Computing on Programmable SoCs via Neural Acceleration&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1701.00485v2 -- &amp;quot;Two-Bit Networks for Deep Learning on Resource-Constrained Embedded Devices&amp;quot; &lt;br /&gt;
&lt;br /&gt;
== Pruning ==&lt;br /&gt;
&lt;br /&gt;
http://papers.nips.cc/paper/5784-learning-both-weights-and-connections-for-efficient-neural-network.pdf --&lt;br /&gt;
&amp;quot;Learning both Weights and Connections for Efficient Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1701.04465 -- &amp;quot;The Incredible Shrinking Neural Network: New Perspectives on Learning Representations Through The Lens of Pruning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1512.08571 -- &amp;quot;Structured Pruning of Deep Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1611.01427 -- &amp;quot;Sparsely-Connected Neural Networks: Towards Efficient VLSI Implementation of Deep Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Efficient Neural Networks via Compression, Quantization, Model Reduction, etc ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1504.04788 -- &amp;quot;Compressing Neural Networks with the Hashing Trick&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1509.08745 -- &amp;quot;Compression of Deep Neural Networks on the Fly&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1502.03436 -- &amp;quot;An Exploration of Parameter Redundancy in Deep Networks with Circulant Projections&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1510.00149 -- &amp;quot;Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.00891 -- &amp;quot;Parameter Compression of Recurrent Neural Networks and Degradation of Short-term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.07061 -- &amp;quot;Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1607.05418 -- &amp;quot;Runtime Configurable Deep Neural Networks for Energy-Accuracy Trade-off&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1602.08194 -- &amp;quot;Scalable and Sustainable Deep Learning via Randomized Hashing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1508.05463 -- &amp;quot;StochasticNet: Forming Deep Neural Networks via Stochastic Connectivity&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1412.7024 -- &amp;quot;Training Deep Neural Networks with Low Precision Multiplications&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1612.03940 -- &amp;quot;Understanding the Impact of Precision Quantization on the Accuracy and Energy of Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1609.00222 -- &amp;quot;Ternary Neural Networks for Resource-Efficient AI Applications&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neural Network Hyperparameter Optimization ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1601.00917 -- &amp;quot;DrMAD: Distilling Reverse-Mode Automatic Differentiation for Optimizing Hyperparameters of Deep Neural Networks&amp;quot; &lt;br /&gt;
&lt;br /&gt;
== Neural Network based EEG Analysis ==&lt;br /&gt;
end&lt;br /&gt;
http://www.seas.upenn.edu/~wulsin/docs/wulsin2010.pdf -- &amp;quot;Semi-Supervised Anomaly Detection for EEG Waveforms Using Deep Belief Nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://inter-eng.upm.ro/2012/files/proceedings/papers/paper72.pdf --&lt;br /&gt;
&amp;quot;Neural Network Parallelization on FPGA Platform for EEG Signal Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Seizure Detection ==&lt;br /&gt;
&lt;br /&gt;
also see https://noisebridge.net/wiki/Kaggle for a (September 2016) current project!&lt;br /&gt;
&lt;br /&gt;
and https://github.com/kevinjos/kaggle-aes-seizure-prediction (some earlier exploration, November 2014)&lt;br /&gt;
&lt;br /&gt;
(broken link, sorry) http://www.sersc.org/journals/ijsip/vol7_no5/26.pdf --&lt;br /&gt;
&amp;quot;A Neural Network Model for Predicting Epileptic Seizures based on Fourier-Bessel Functions&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2230664/pdf/nihms-34528.pdf --&lt;br /&gt;
&amp;quot;A Stochastic Framework for Evaluating Seizure Prediction Algorithms Using Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
(another broken link) http://cs.uni-muenster.de/Professoren/Lippe/diplomarbeiten/html/eisenbach/Untersuchte%20Artikel/PPHD+00.pdf --&lt;br /&gt;
&amp;quot;Recurrent neural network based preenddiction of epileptic seizures in intra- and extracranial EEG&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Visible Light Sensor Network ==&lt;br /&gt;
&lt;br /&gt;
http://infoteh.etf.unssa.rs.ba/zbornik/2016/radovi/KST-1/KST-1-15.pdf --&lt;br /&gt;
&amp;quot;Analysis of Visible Light Communication System for Implementation in Sensor Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neurophysiology ==&lt;br /&gt;
&lt;br /&gt;
http://www.buzsakilab.com/content/PDFs/BuzsakiKoch2012.pdf -- &amp;quot;The origin of extracellular fields and&lt;br /&gt;
currents — EEG, ECoG, LFP and spikes&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Signal Processing ==&lt;br /&gt;
&lt;br /&gt;
http://provideyourown.com/2011/analogwrite-convert-pwm-to-voltage/ -- &amp;quot;Arduino’s AnalogWrite – Converting PWM to a Voltage&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://sim.okawa-denshi.jp/en/PWMtool.php -- &amp;quot;RC Low-pass Filter Design for PWM (Transient Analysis Calculator)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Hyperdimensional Computing ==&lt;br /&gt;
&lt;br /&gt;
http://redwood.berkeley.edu/pkanerva/papers/kanerva09-hyperdimensional.pdf --&lt;br /&gt;
&amp;quot;Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1602.03032.pdf --&lt;br /&gt;
&amp;quot;Associative Long Short-Term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Bird Flocks and Maximum Entropy ==&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1107.0604v1 --&lt;br /&gt;
&amp;quot;Statistical Mechanics and Flocks of Birds&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1307.5563v1 --&lt;br /&gt;
&amp;quot;Social interactions dominate speed control in driving natural flocks toward criticality&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ocw.mit.edu/courses/electrical-engineering-and-computer-science/6-050j-information-and-entropy-spring-2008/syllabus/MIT6_050JS08_penfield.pdf --&lt;br /&gt;
&amp;quot;Information and Entropy (Course Notes)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Whale Songs ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1307.0589.pdf -- &amp;quot;The Orchive : Data mining a massive bioacoustic archive&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.researchgate.net/profile/Herbert_Roitblat/publication/13429327_The_neural_network_classification_of_false_killer_whale_%28Pseudorca_crassidens%29_vocalizations/links/540d2ff60cf2df04e75478cd.pdf -- &amp;quot;The neural network classification of false killer whale (Pseudorca crassidens) vocalizations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://users.iit.demokritos.gr/~paliourg/papers/PhD.pdf -- &amp;quot;REFINEMENT OF TEMPORAL CONSTRAINTS IN AN EVENT RECOGNITION SYSTEM USING SMALL DATASETS&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.nersc.no/sites/www.nersc.no/files/master_thesis_sebastian_menze.pdf -- &amp;quot;Estimating fin whale distribution from ambient noise spectra using Bayesian inversion&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://sis.univ-tln.fr/~glotin/IJCNN2015_IHMMbioac_BartChamGlot.pdf -- &amp;quot;Hierarchical Dirichlet Process Hidden Markov Model for Unsupervised Bioacoustic Analysis&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.inf.ed.ac.uk/publications/thesis/online/IM030057.pdf -- &amp;quot;Hidden Markov Model Clustering of Acoustic Data&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://danielnouri.org/notes/2014/01/10/using-deep-learning-to-listen-for-whales/ -- Using deep learning to listen for whales&lt;br /&gt;
&lt;br /&gt;
== Computational Cognitive Neuroscience ==&lt;br /&gt;
&lt;br /&gt;
http://www.pnas.org/content/110/41/16390.full -- &amp;quot;Indirection and symbol-like processing in the prefrontal cortex and basal ganglia&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ruccs.rutgers.edu/images/personal-zenon-pylyshyn/docs/jaf.pdf -- &amp;quot;Connectionism and Cognitive Architecture: A Critical Analysis&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2728678/pdf/nihms131814.pdf -- Neves et al 2008 &amp;quot;Cell Shape and Negative Links in Regulatory Motifs Together Control Spatial Information Flow in Signaling Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://psych.colorado.edu/~oreilly/papers/AisaMingusOReilly08.pdf -- &amp;quot;The Emergent Neural Modeling System&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://grey.colorado.edu/emergent/index.php/Comparison_of_Neural_Network_Simulators&lt;br /&gt;
&lt;br /&gt;
== Text Generation ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~ilya/pubs/2011/LANG-RNN.pdf -- &amp;quot;Generating Text with Recurrent Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Games ==&lt;br /&gt;
&lt;br /&gt;
http://setgame.com/sites/default/files/teacherscorner/COGNITIVE%20MODELING%20WITH%20SET.pdf -- &amp;quot;How to Construct a Believable Opponent using Cognitive Modeling in the Game of Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www-personal.umich.edu/~charchan/SET.pdf -- &amp;quot;SETs and Anti-SETs: The Math Behind the Game of SET&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://personal.plattsburgh.edu/quenelgt/talks/set.pdf -- &amp;quot;Introduction to Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://web.engr.illinois.edu/~pbg/papers/set.pdf -- &amp;quot;On the Complexity of the Game of Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.warwick.ac.uk/staff/D.Maclagan/papers/set.pdf -- &amp;quot;The Card Game Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.math.ucdavis.edu/~anne/FQ2011/set_game.pdf -- &amp;quot;The Game Set&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Large Scale Brain Simulation ==&lt;br /&gt;
&lt;br /&gt;
http://www.nowere.net/b/arch/96550/src/1378907656268.pdf -- &amp;quot;A world survey of artificial brain projects, Part I: Large-scale brain simulations&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Music ==&lt;br /&gt;
&lt;br /&gt;
http://cmr.soc.plymouth.ac.uk/publications/bci-wkshop.pdf -- &amp;quot;ON GENERATING EEG FOR CONTROLLING MUSICAL SYSTEMS&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
https://github.com/nbdt/gotrain (our ANN code)&lt;br /&gt;
&lt;br /&gt;
https://github.com/nbdt/openbci-golang-server (EEG data acquisition and visualization software for OpenBCI)&lt;br /&gt;
&lt;br /&gt;
== Hidden Markov Models ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.sjsu.edu/~stamp/RUA/HMM.pdf&lt;br /&gt;
-- &amp;quot;A Revealing Introduction to Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.jelmerborst.nl/pubs/Borst2013b.pdf&lt;br /&gt;
-- &amp;quot;Discovering Processing Stages by combining EEG with Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2230664/pdf/nihms-34528.pdf&lt;br /&gt;
-- &amp;quot;A Stochastic Framework for Evaluating Seizure Prediction Algorithms Using Hidden Markov Models&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.schalklab.org/sites/default/files/misc/Coupled%20hidden%20markov%20model%20for%20electrocorticographic%20signal%20classification.pdf&lt;br /&gt;
-- &amp;quot;Coupled Hidden Markov Model for Electrocorticographic Signal Classification&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Long Short Term Memory ==&lt;br /&gt;
&lt;br /&gt;
http://deeplearning.cs.cmu.edu/pdfs/Hochreiter97_lstm.pdf&lt;br /&gt;
-- &amp;quot;Long Short-Term Memory&amp;quot;&lt;br /&gt;
&lt;br /&gt;
ftp://ftp.idsia.ch/pub/juergen/2002_icannMusic.pdf&lt;br /&gt;
-- &amp;quot;Learning The Long-Term Structure of the Blues&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.overcomplete.net/papers/nn2012.pdf&lt;br /&gt;
-- &amp;quot;A generalized LSTM-like training algorithm for second-order recurrent neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.iro.umontreal.ca/~eckdoug/papers/2003_nn.pdf&lt;br /&gt;
-- &amp;quot;Kalman filters improve LSTM network performance in problems unsolvable by traditional recurrent nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://googleresearch.blogspot.ch/2015/09/google-voice-search-faster-and-more.html&lt;br /&gt;
-- &amp;quot;Long Short-Term Memory dramatically improves Google Voice etc&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1511.05552v4.pdf --&lt;br /&gt;
&amp;quot;Recurrent Neural Networks Hardware Implementation on FPGA&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://vast.cs.ucla.edu/sites/default/files/publications/ASP-DAC2017-1352-11.pdf --&lt;br /&gt;
&amp;quot;FPGA-based Accelerator for Long Short-Term Memory Recurrent Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Question Answering ==&lt;br /&gt;
&lt;br /&gt;
http://www.overcomplete.net/papers/bica2012.pdf&lt;br /&gt;
-- &amp;quot;Neural Architectures for Learning to Answer Questions&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://cs.umd.edu/~miyyer/pubs/2014_qb_rnn.pdf&lt;br /&gt;
-- &amp;quot;A Neural Network for Factoid Question Answering over Paragraphs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1502.05698.pdf&lt;br /&gt;
-- &amp;quot;Towards AI-Complete Question Answering: A Set of Prerequisite Toy Tasks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.peterschueller.com/pub/2014/2014_winograd_schemas_relevance_knowledge_graphs.pdf&lt;br /&gt;
-- &amp;quot;Tackling Winograd Schemas by Formalizing Relevance Theory in Knowledge Graphs&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://ijcai.org/papers15/Papers/IJCAI15-190.pdf&lt;br /&gt;
-- &amp;quot;Towards Addressing the Winograd Schema Challenge — Building and Using a Semantic Parser and a Knowledge Hunting Module&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1506.05869v2.pdf&lt;br /&gt;
-- &amp;quot;A Neural Conversational Model&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1508.05508v1.pdf&lt;br /&gt;
-- &amp;quot;Towards Neural Network-based Reasoning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.visualqa.org/vqa_iccv2015.pdf&lt;br /&gt;
-- &amp;quot;VQA: Visual Question Answering&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Propagators ==&lt;br /&gt;
Cells must support three operations:&lt;br /&gt;
*add some content&lt;br /&gt;
*collect the content currently accumulated&lt;br /&gt;
*register a propagator to be notified when the accumulated content changes&lt;br /&gt;
*When new content is added to a cell, the cell must merge the addition with the content already present. When a propagator asks for the content of a cell, the cell must deliver a complete summary of the information that has been added to it.&lt;br /&gt;
*The merging of content must be commutative, associative, and idempotent. The behavior of propagators must be monotonic with respect to the lattice induced by the merge operation.&lt;br /&gt;
*http://web.mit.edu/~axch/www/art.pdf *http://groups.csail.mit.edu/mac/users/gjs/propagators/&lt;br /&gt;
*http://dustycloud.org/blog/sussman-on-ai/&lt;br /&gt;
&lt;br /&gt;
== Boosting ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.princeton.edu/courses/archive/spr08/cos424/readings/Schapire2003.pdf&lt;br /&gt;
-- &amp;quot;The Boosting Approach to Machine Learning An Overview&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.19.9955&amp;amp;rep=rep1&amp;amp;type=pdf&lt;br /&gt;
-- &amp;quot;Ensembling Neural Networks: Many Could Be Better Than All&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cs.columbia.edu/~rocco/Public/mlj9.pdf&lt;br /&gt;
-- &amp;quot;Random Classification Noise Defeats All Convex Potential Boosters&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Support Vector Machines ==&lt;br /&gt;
&lt;br /&gt;
http://research.microsoft.com/en-us/um/people/cburges/papers/svmtutorial.pdf&lt;br /&gt;
-- &amp;quot;A Tutorial on Support Vector Machines for Pattern Recognition&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Wire Length / Small World Networks ==&lt;br /&gt;
&lt;br /&gt;
http://koulakovlab.cshl.edu/publications/koulakov_chklovskii_2000.pdf&lt;br /&gt;
-- &amp;quot;A wire length minimization approach to ocular dominance patterns in mammalian visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://papers.nips.cc/paper/1910-foundations-for-a-circuit-complexity-theory-of-sensory-processing.pdf&lt;br /&gt;
-- &amp;quot;Foundations for a Circuit Complexity Theory of Sensory Processing&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.nada.kth.se/~cjo/documents/small_world.pdf&lt;br /&gt;
-- &amp;quot;Small-World Connectivity and Attractor Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.doc.ic.ac.uk/~mpsha/ShanahanPhysRevEPreprint.pdf&lt;br /&gt;
-- &amp;quot;The Dynamical Complexity of Small-World Networks of Spiking Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dam.brown.edu/people/elie/papers/small_world.pdf&lt;br /&gt;
-- &amp;quot;Autoassociative Memory Retrieval and Spontaneous Activity Bumps in Small-World Networks of Integrate-and-Fire Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.332.4501&amp;amp;rep=rep1&amp;amp;type=pdf&lt;br /&gt;
-- &amp;quot;Transition from Random to Small-World Neural Networks by STDP Learning Rule&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://tamar.tau.ac.il/~eshel/papers/J_Neural_Engineering.pdf&lt;br /&gt;
-- &amp;quot;Compact self-wiring in cultured neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Backpropagation ==&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/neuron.pdf -- &amp;quot;Neural Networks - A Systematic Introduction&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www4.rgu.ac.uk/files/chapter3%20-%20bp.pdf (works through simple example)&lt;br /&gt;
&lt;br /&gt;
http://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example/&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/chapter/K7.pdf&lt;br /&gt;
&lt;br /&gt;
also see http://page.mi.fu-berlin.de/rojas/neural/neuron.pdf (entire book &amp;quot;Neural Networks - a Systemic Introduction&amp;quot; by Raul Rojas)&lt;br /&gt;
&lt;br /&gt;
http://work.caltech.edu/lectures.html Hoeffding&#039;s inequality, VC Dimension and Back Propagation ANN&lt;br /&gt;
&lt;br /&gt;
http://www.cs.stir.ac.uk/research/publications/techreps/pdf/TR148.pdf (&amp;quot;Learning XOR: exploring the space of a classic problem&amp;quot;)&lt;br /&gt;
&lt;br /&gt;
http://www.sysc.pdx.edu/classes/Werbos-Backpropagation%20through%20time.pdf&lt;br /&gt;
-- &amp;quot;Backpropagation Through Time: What it Does and How to Do It&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Convolutional Neural Networks ==&lt;br /&gt;
&lt;br /&gt;
http://scs.ryerson.ca/~aharley/vis/harley_vis_isvc15.pdf -- &amp;quot;An Interactive Node-Link Visualization&lt;br /&gt;
of Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://white.stanford.edu/teach/index.php/An_Introduction_to_Convolutional_Neural_Networks&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Dosovitskiy_Learning_to_Generate_2015_CVPR_paper.pdf -- &amp;quot;Learning to Generate Chairs With Convolutional Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.pamitc.org/cvpr15/files/lecun-20150610-cvpr-keynote.pdf&lt;br /&gt;
-- &amp;quot;What&#039;s Wrong With Deep Learning?&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Computer Vision ==&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/CVPR2015.py -- some interesting papers here&lt;br /&gt;
&lt;br /&gt;
http://www.cv-foundation.org/openaccess/content_cvpr_2015/papers/Nguyen_Deep_Neural_Networks_2015_CVPR_paper.pdf -- &amp;quot;Deep Neural Networks are Easily Fooled: High Confidence Predictions for Unrecognizable Images&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Visual Perception (Biological Systems) ==&lt;br /&gt;
&lt;br /&gt;
http://cbcl.mit.edu/publications/ps/Serre_etal_PBR07.pdf&lt;br /&gt;
-- &amp;quot;A quantitative theory of immediate visual recognition&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dam.brown.edu/ptg/REPORTS/Invariance.pdf&lt;br /&gt;
-- &amp;quot;Invariance and Selectivity in the Ventral Visual Pathway&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cnbc.cmu.edu/~tai/papers/lee_mumford_josa.pdf&lt;br /&gt;
-- &amp;quot;Hierarchical Bayesian inference in the visual cortex&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Neural Synchrony ==&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1312.6115.pdf&lt;br /&gt;
-- &amp;quot;Neuronal Synchrony in Complex-Valued Deep Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
== Spiking Neural Networks ==&lt;br /&gt;
&lt;br /&gt;
http://ncs.ethz.ch/projects/evospike/publications/ICONIP2011%20Springer%20LNCS%20Nutta.pdf -- &amp;quot;EEG Classification with BSA Spike Encoding Algorithm and Evolving Probabilistic Spiking Neural Network&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://dai.fmph.uniba.sk/~benus/confers/CNGM_ICANN05.pdf -- &amp;quot;Computational Neurogenetic Modeling: Integration of Spiking Neural Networks, Gene Networks, and Signal Processing Techniques&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://citeseerx.ist.psu.edu/viewdoc/summary?doi=10.1.1.97.3902 -- &amp;quot;Pattern Recognition in a Bucket&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.igi.tugraz.at/maass/psfiles/221.pdf -- &amp;quot;Noise as a Resource for Computation and Learning in Spiking Neural Networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://web.stanford.edu/group/brainsinsilicon/goals.html -- Neurogrid&lt;br /&gt;
&lt;br /&gt;
http://apt.cs.manchester.ac.uk/projects/SpiNNaker/ -- SpiNNaker&lt;br /&gt;
&lt;br /&gt;
http://personalpages.manchester.ac.uk/staff/javier.navaridas/pubs/2012.jpdc.pdf -- more on SpiNNaker&lt;br /&gt;
&lt;br /&gt;
==Hierarchical Temporal Memory==&lt;br /&gt;
&lt;br /&gt;
https://github.com/numenta/nupic/wiki/Hierarchical-Temporal-Memory-Theory&lt;br /&gt;
&lt;br /&gt;
http://journals.plos.org/ploscompbiol/article?id=10.1371/journal.pcbi.1000532 -- &amp;quot;Towards a Mathematical Theory of Cortical Micro-circuits&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Distributed Neural Networks==&lt;br /&gt;
&lt;br /&gt;
https://www.paypal-engineering.com/2015/01/12/deep-learning-on-hadoop-2-0-2/ -- paypal engineering blog, paper describes implementation of neural network as described by [http://www.maths.bris.ac.uk/~maxvd/reduce_dim.pdf Hinton 2006] on Hadoop&lt;br /&gt;
&lt;br /&gt;
http://www.sciencemag.org/content/suppl/2006/08/04/313.5786.504.DC1/Hinton.SOM.pdf -- supplementary material for [http://www.maths.bris.ac.uk/~maxvd/reduce_dim.pdf Hinton 2006]&lt;br /&gt;
&lt;br /&gt;
http://www.cs.otago.ac.nz/staffpriv/hzy/papers/pdcs03.pdf -- &amp;quot;Parallelization of a Backpropagation Neural Network on a Cluster Computer&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1404.5997v2.pdf -- &amp;quot;One weird trick for parallelizing convolutional neural networks&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.ics.uci.edu/~jmoorkan/pub/gpusnn-ijcnn.pdf -- &amp;quot;Efficient Simulation of Large-Scale Spiking Neural Networks Using CUDA Graphics Processors&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Hopfield nets and RBMs==&lt;br /&gt;
&lt;br /&gt;
http://www.iro.umontreal.ca/~bengioy/papers/ftml_book.pdf -- Yoshua Bengio, 2009, Learning Deep Architectures for AI&lt;br /&gt;
&lt;br /&gt;
http://deeplearning.cs.cmu.edu/ -- Syllabus for cs course on deep learning, possible source of literature for the library &lt;br /&gt;
&lt;br /&gt;
https://class.coursera.org/neuralnets-2012-001/lecture -- Lecture 11 - 15 to aid in understanding of Hinton 2006 paper above&lt;br /&gt;
&lt;br /&gt;
http://www.pnas.org/content/79/8/2554.full.pdf -- &amp;quot;Neural networks and physical systems with emergent collective computational abilities&amp;quot; (Hopfield 1982)&lt;br /&gt;
&lt;br /&gt;
http://page.mi.fu-berlin.de/rojas/neural/chapter/K13.pdf -- &amp;quot;The Hopfield Model&amp;quot; (Rojas 1996)&lt;br /&gt;
&lt;br /&gt;
http://www.seas.upenn.edu/~wulsin/docs/wulsin2010.pdf -- &amp;quot;Semi-Supervised Anomaly Detection for EEG Waveforms Using Deep Belief Nets&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cse.buffalo.edu/DBGROUP/bioinformatics/papers/cameraready_xwjiabibe.pdf -- &amp;quot;A Novel Semi-supervised Deep Learning Framework&lt;br /&gt;
for Affective State Recognition on EEG Signals&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~hinton/absps/guideTR.pdf -- &amp;quot;A Practical Guide to Training Restricted Boltzmann Machines&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1503.07793v2.pdf&lt;br /&gt;
-- &amp;quot;Gibbs Sampling with Low-Power Spiking Digital Neurons&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://arxiv.org/pdf/1311.0190v1 -- &amp;quot;On the typical properties of inverse problems in statistical mechanics&amp;quot; Iacopo Mastromatteo 2013&lt;br /&gt;
&lt;br /&gt;
http://jmlr.org/proceedings/papers/v5/salakhutdinov09a/salakhutdinov09a.pdf -- &amp;quot;Deep Boltzmann Machines&amp;quot; Salakhutdinov &amp;amp; Hinton 2009&lt;br /&gt;
&lt;br /&gt;
http://www.cs.toronto.edu/~hinton/absps/tr00-004.pdf -- &amp;quot;Training Products of Experts by Minimizing Contrastive Divergence&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.eecg.toronto.edu/~pc/research/publications/ly.fpga2009.submitted.pdf -- &amp;quot;A High-Performance FPGA Architecture for Restricted&lt;br /&gt;
Boltzmann Machines&amp;quot; Ly &amp;amp; Chow 2009&lt;br /&gt;
&lt;br /&gt;
https://pdfs.semanticscholar.org/85fa/f7c3c05388e2bcd097a416606bdd88fc0c7c.pdf -- &amp;quot;A MULTI-FPGA ARCHITECTURE FOR STOCHASTIC RESTRICTED BOLTZMANN MACHINES&amp;quot; Ly &amp;amp; Chow 2009&lt;br /&gt;
&lt;br /&gt;
== Neuromorphic Stuff ==&lt;br /&gt;
&lt;br /&gt;
https://arxiv.org/pdf/1508.01008.pdf -- &amp;quot;INsight: A Neuromorphic Computing System for Evaluation of Large Neural Networks&amp;quot; Chung, Shin &amp;amp; Kang 2015&lt;br /&gt;
&lt;br /&gt;
== Markov Chain Monte Carlo ==&lt;br /&gt;
&lt;br /&gt;
http://www.cs.princeton.edu/courses/archive/spr06/cos598C/papers/AndrieuFreitasDoucetJordan2003.pdf&lt;br /&gt;
-- &amp;quot;An Introduction to MCMC for Machine Learning&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://jmlr.org/proceedings/papers/v37/salimans15.pdf&lt;br /&gt;
-- &amp;quot;Markov Chain Monte Carlo and Variational Inference: Bridging the Gap&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.umiacs.umd.edu/~resnik/pubs/LAMP-TR-153.pdf -- &amp;quot;Gibbs Sampling for the Uninitiated&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Entrainment==&lt;br /&gt;
&lt;br /&gt;
http://www.scirp.org/Journal/PaperDownload.aspx?paperID=33251 -- &amp;quot;Alpha Rhythms Response to 10 Hz Flicker Is Wavelength Dependent&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.brainmachine.co.uk/wp-content/uploads/Herrmann_Flicker.pdf -- &amp;quot;EEG responses to 1–100 Hz flicker: resonance phenomena in visual cortex and their potential correlation to cognitive phenomena&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.jneurosci.org/content/23/37/11621.full.pdf -- &amp;quot;Human Cerebral Activation during Steady-State Visual-Evoked Responses&amp;quot;&lt;br /&gt;
&lt;br /&gt;
http://www.dauwels.com/Papers/CogDyn%202009.pdf -- &amp;quot;On the synchrony of steady state visual evoked potentials and oscillatory burst events&amp;quot;&lt;br /&gt;
&lt;br /&gt;
https://www.tu-ilmenau.de/fileadmin/public/lorentz-force/publications/peer/2012/haueisen2012/Halbleib_JCN_2012_Topographic_analysis_photic_driving.pdf -- &amp;quot;Topographic Analysis of Engagement and Disengagement of Neural Oscillators in Photic Driving: A Combined Electroencephalogram/Magnetoencephalogram Study&amp;quot;&lt;br /&gt;
&lt;br /&gt;
==Mining Scientific Literature==&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC153503/pdf/1471-2105-4-11.pdf -- &amp;quot;PreBIND and Textomy – mining the biomedical literature for protein-protein interactions using a support vector machine&amp;quot; Donaldson 2003 BMC Bioinformatics&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4674139/pdf/pcbi.1004630.pdf -- &amp;quot;Text Mining for Protein Docking&amp;quot; Badal 2015 PLoS Comput Biol.&lt;br /&gt;
&lt;br /&gt;
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4691339/pdf/bav116.pdf -- &amp;quot;Biocuration with insufficient resources and fixed timelines&amp;quot; Rodriguez-Esteban 2015 Database: The Journal of Biological Databases and Curation&lt;br /&gt;
&lt;br /&gt;
==(not necessarilly very) Current Discussion==&lt;br /&gt;
&lt;br /&gt;
re Tononi&#039;s &amp;quot;Integrated Information Theory&amp;quot; http://www.scottaaronson.com/blog/?p=1799&lt;br /&gt;
&lt;br /&gt;
(19 February 2014) starting to think about possibility for experiments (loosely) related to [https://en.wikipedia.org/wiki/Visual_evoked_potential Visual Evoked Potential] research again - for instance:&lt;br /&gt;
&lt;br /&gt;
http://www.biosemi.com/publications/pdf/Bierman.pdf -- Wave function collapse&lt;br /&gt;
&lt;br /&gt;
[[File:Schak99InstantaneousCoherenceStroopTask.pdf]] -- &amp;quot;Instantaneous EEG Coherence Analysis During the Stroop Task&amp;quot; -- Schack et al 1999&lt;br /&gt;
&lt;br /&gt;
[[File:CoherentEEGAmbiguousFigureBinding.pdf]] -- &amp;quot;Coherent EEG Indicators of Cognitive Binding During Ambiguous Figure Tasks&amp;quot; -- Klemm, Li, and Hernandez 2000 &lt;br /&gt;
&lt;br /&gt;
Note these two papers flog coherence measures - not trying to focus so much on that analysis right now, more interested in general understanding of what these experiments are about with possible goal of designing simpler experiments &amp;amp; analysis of similar perceptual/cognitive phenomena.&lt;br /&gt;
&lt;br /&gt;
Here is an article that looks more directly at visual evoked potential measures:&lt;br /&gt;
&lt;br /&gt;
[[File:ERP_Stereoscopic.pdf]] -- &amp;quot;Evoked Responses to Distinct and Nebulous Stereoscopic Stimuli&amp;quot; -- Dunlop et al 1983&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
(11 September 2013) more on analysis methods:&lt;br /&gt;
&lt;br /&gt;
http://slesinsky.org/brian/misc/eulers_identity.html&lt;br /&gt;
&lt;br /&gt;
http://www.dspguide.com/ch8/1.htm&lt;br /&gt;
&lt;br /&gt;
[[File:Fftw3.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ParametricEEGAnalysis.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ICATutorial.pdf]]&lt;br /&gt;
&lt;br /&gt;
[[File:ICAFrequencyDomainEEG.pdf]]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
(21 August 2013) - readings relating statistical (etc math / signal processing / pattern recognition / machine learning) methods for EEG data interpretation.  A lot of stuff, a bit of nonsense ... and ... statistics!&lt;br /&gt;
&lt;br /&gt;
Would be good to identify any papers suitable for more in-depth study.  Currently have a wide field to graze for selections:&lt;br /&gt;
&lt;br /&gt;
[[File:DWTandFFTforEEG.pdf]] &amp;quot;EEG Classifier using Fourier Transform and Wavelet Transform&amp;quot; -- Maan Shaker, 2007&lt;br /&gt;
&lt;br /&gt;
[[File:Schak99InstantaneousCoherenceStroopTask.pdf]] -- &amp;quot;Instantaneous EEG Coherence Analysis During the Stroop Task&amp;quot; -- Schack et al 1999&lt;br /&gt;
&lt;br /&gt;
[[File:KulaichevCoherence.pdf]] -- &amp;quot;The Informativeness of Coherence Analysis in EEG Studies&amp;quot; -- A. P. Kulaichev 2009 &#039;&#039;note: interesting critical perspective re limitations, discussion of alternative analytics&#039;&#039;&lt;br /&gt;
&lt;br /&gt;
[[File:ContinuousAndDiscreteWaveletTransforms.pdf]] -- review of (pre-1990) wavelet literature -- Christopher Heil and David Walnut, 1989&lt;br /&gt;
&lt;br /&gt;
[[File:EEGGammaMeditation.pdf]] -- &amp;quot;Brain sources of EEG gamma frequency during volitionally meditation-induced, altered states of consciousness, and experience of the self&amp;quot; -- Dietrich Lehman et al 2001&lt;br /&gt;
&lt;br /&gt;
http://neuro.hut.fi/~pavan/home/Hyvarinen2010_FourierICA_Neuroimage.pdf - &amp;quot;Independent component analysis of short-time Fourier transforms for spontaneous EEG/MEG analysis&amp;quot; -- Aapo Hyvarinen, Pavan Ramkumar, Lauri Parkkonen, Riitta Hari - paper published in Neuroimage vol 49 (2010)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://www.nickgillian.com/software/grt OpenSource Machine Learning Algs from NG @MIT]&lt;br /&gt;
&amp;lt;br&amp;gt;[https://www.usenix.org/system/files/conference/usenixsecurity12/sec12-final56.pdf Consumer grade EEG used to see &amp;quot;P300&amp;quot; reponse] and for thoes with a short attention span [http://www.extremetech.com/extreme/134682-hackers-backdoor-the-human-brain-successfully-extract-sensitive-data tldr]&lt;br /&gt;
&amp;lt;br&amp;gt;(discussed at meetup Wednesday 31 July 2013)&lt;br /&gt;
&amp;lt;br&amp;gt;&amp;quot;Coherent EEG Indicators of Cognitive Binding During Ambiguous Figure Tasks&amp;quot; Klemm, Li, and Hernandez 2000 &lt;br /&gt;
&amp;lt;br&amp;gt;[[File:CoherentEEGAmbiguousFigureBinding.pdf]]&lt;br /&gt;
&amp;lt;br&amp;gt;&amp;quot;We tested the hypothesis that perception of an alternative image in ambiguous figures would be manifest as high-frequency (gamma) components that become synchronized over multiple scalp sites as a &amp;quot;cognitive binding&amp;quot; process occurs.&amp;quot;&lt;br /&gt;
&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://dreamsessions.net art, dream, and eeg]&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[http://www.believermag.com/issues/200710/?read=article_aviv mind v brain, hobson v solms]&lt;br /&gt;
&amp;lt;br&amp;gt;http://en.wikipedia.org/wiki/Activation-synthesis_hypothesis&lt;br /&gt;
&amp;lt;br&amp;gt;[[File:HobsomREMDreamProtoconsciousness.pdf|Hobson09ProtosconsciousnessREMDream]]&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Hobson and McCarley originally proposed in the 1970s that the differences in the waking-NREM-REM sleep cycle was the result of interactions between aminergic REM-off cells and cholinergic REM-on cells.[4] This was perceived as the activation-synthesis model, stating that brain activation during REM sleep results in synthesis of dream creation.[1][1] Hobson&#039;s five cardinal characteristics include: intense emotions, illogical content, apparent sensory impressions, uncritical acceptance of dream events, and difficulty in being remembered.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
Berkeley Labs&lt;br /&gt;
&lt;br /&gt;
[http://gallantlab.org/index.html Gallant Group]&lt;br /&gt;
&amp;lt;br&amp;gt;[http://walkerlab.berkeley.edu/ Walker Group]&lt;br /&gt;
&amp;lt;br&amp;gt;[http://socrates.berkeley.edu/~plab/ Palmer Group]&lt;br /&gt;
&lt;br /&gt;
==Sleep Research==&lt;br /&gt;
&lt;br /&gt;
[http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2335403/ Comment on the AASM Manual for the Scoring of Sleep and Associated Events]&lt;br /&gt;
&lt;br /&gt;
==random tangents==&lt;br /&gt;
(following previous discussion) - we might select a few to study in more depth&lt;br /&gt;
(... or not!  Plenty more to explore - suggestions (random or otherwise) are welcome.&lt;br /&gt;
http://www.meltingasphalt.com/neurons-gone-wild/ --&lt;br /&gt;
Neurons Gone Wild - Levels of agency in the brain. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;stereoscopic perception:&#039;&#039;&#039;&lt;br /&gt;
*[[File:ERP_Stereoscopic.pdf]] &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
some (maybe) interesting background on Information Theory (cool title...)&lt;br /&gt;
 Claude Shannon: &amp;quot;Communication in the Presence of Noise&amp;quot;&lt;br /&gt;
 [[File:Shannon_noise.pdf]]&lt;br /&gt;
 &amp;quot;We will call a system that transmits without errors at the rate &#039;&#039;C&#039;&#039; an ideal system.&lt;br /&gt;
  Such a system cannot be achieved with any finite encoding process&lt;br /&gt;
  but can be approximated as closely as desired.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
wikipedia etc quick reads:&lt;br /&gt;
 https://en.wikipedia.org/wiki/Eeg&lt;br /&gt;
 https://en.wikipedia.org/wiki/Neural_synchronization&lt;br /&gt;
 https://en.wikipedia.org/wiki/Event-related_potentials&lt;br /&gt;
 http://www.scholarpedia.org/article/Spike-and-wave_oscillations&lt;br /&gt;
 http://www.scholarpedia.org/article/Thalamocortical_oscillations&lt;br /&gt;
&lt;br /&gt;
==Previously==&lt;br /&gt;
&lt;br /&gt;
[http://www.psychiclab.net/ Masahiro&#039;s EEG Device/IBVA Software]&lt;br /&gt;
&lt;br /&gt;
[http://www.instructables.com/id/open-brain-wave-interface-hardware-1/ and ... open source hardware design and kits on instructables.com]&lt;br /&gt;
&lt;br /&gt;
[http://brainstorms.puzzlebox.info/ Puzzlebox - Opensource BCI Developers]&lt;br /&gt;
&lt;br /&gt;
Morgan from GazzLab @ MissionBay/UCSF&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
https://github.com/gazzlab&lt;br /&gt;
&lt;br /&gt;
Let&#039;s ease into a lightweight &amp;quot;journal club&amp;quot; discussion with this technical report from NeuroSky.&lt;br /&gt;
&lt;br /&gt;
Name: A user-friendly SSVEP-based brain-computer interface using a time-domain classifier, Luo A and Sullivan TJ 2010&lt;br /&gt;
&lt;br /&gt;
URL: [[File:NeuroSkyVEP.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
Please add your comments &amp;amp; questions here.&lt;br /&gt;
&lt;br /&gt;
==Background Reading==&lt;br /&gt;
&lt;br /&gt;
http://nanosouffle.net/ (view into Arxiv.org)&lt;br /&gt;
&lt;br /&gt;
Name: Hunting for Meaning after Midnight, Miller 2007&lt;br /&gt;
&lt;br /&gt;
URL: &amp;lt;http://www.noisebridge.net/pipermail/neuro/attachments/20130501/4b992eb6/attachment-0002.pdf&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Name: Broken mirrors, Ram, VS, &amp;amp; Oberman, LM, 2006, Nov&lt;br /&gt;
&lt;br /&gt;
URL: &amp;lt;http://www.noisebridge.net/pipermail/neuro/attachments/20130501/4b992eb6/attachment-0003.pdf&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Ramachandran Critique&lt;br /&gt;
&lt;br /&gt;
http://blogs.scientificamerican.com/guest-blog/2012/11/06/whats-so-special-about-mirror-neurons/&lt;br /&gt;
&lt;br /&gt;
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2773693/&lt;br /&gt;
&lt;br /&gt;
Sleep/Dream Studies&lt;br /&gt;
&lt;br /&gt;
http://www.cns.atr.jp/dni/en/publications/&lt;br /&gt;
&lt;br /&gt;
==NeuroSky Docs==&lt;br /&gt;
[[File:NeuroSkyDongleProtocol.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
[[File:NeuroSkyCommunicationsProtocol.pdf‎]]&lt;br /&gt;
&lt;br /&gt;
==Android Neutral Network Fuzzy Learning app==&lt;br /&gt;
[https://play.google.com/store/apps/details?id=com.faadooengineers.free_neuralnetworkandfuzzysystems Android Neutral Network Fuzzy Learning app in Play Store]&lt;br /&gt;
&lt;br /&gt;
==Learning about Neural Networks==&lt;br /&gt;
* What type of network? [http://stats.stackexchange.com/questions/114385/what-is-the-difference-between-convolutional-neural-networks-restricted-boltzma RMB (Restricted Boltzmann Machine) vs Autoencoder/MLP vs CNN (Convolutional Neural Networks)]&lt;br /&gt;
* Andrej Karpathy&#039;s [http://cs.stanford.edu/people/karpathy/convnetjs/ Convolutional Neural Network coded in JavaScript (ConvNetJS)]&lt;br /&gt;
* Andrej Karpathy&#039;s [http://karpathy.github.io/2015/10/25/selfie/ What a Deep Neural Network thinks about your #selfie  (background on Convolutional Neural Networks for image recognition and classification)]&lt;br /&gt;
* [https://blog.webkid.io/neural-networks-in-javascript/ Neural Networks in JavaScript w/MNIST]&lt;br /&gt;
* [http://www.antoniodeluca.info/blog/10-08-2016/neural-networks-in-javascript.html Another NN in JS]&lt;br /&gt;
* [http://caza.la/synaptic/ The Synaptic &amp;quot;architecture-free&amp;quot; neural network library in JS]&lt;/div&gt;</summary>
		<author><name>198.27.140.180</name></author>
	</entry>
	<entry>
		<id>https://wiki.extremist.software/index.php?title=Talk:DreamTeam/Brainduino&amp;diff=57402</id>
		<title>Talk:DreamTeam/Brainduino</title>
		<link rel="alternate" type="text/html" href="https://wiki.extremist.software/index.php?title=Talk:DreamTeam/Brainduino&amp;diff=57402"/>
		<updated>2017-03-22T03:09:33Z</updated>

		<summary type="html">&lt;p&gt;198.27.140.180: /* March 21 2017 */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==March 21 2017==&lt;br /&gt;
Q&amp;amp;A with Masahiro!&lt;br /&gt;
&lt;br /&gt;
&amp;lt;&amp;lt;&lt;br /&gt;
We are making progress on our Brainduino V0.1 assembly. We think we isolated the abnormality to IC3, the low-pass filters. We observe 10,000x gain out of the first two rounds of amplification. Our understanding is that 10,000x gain amplifies a +/-50uV brain powered signal to a +/-0.5V signal, within the operational range of the Arduino ACD.&lt;br /&gt;
&amp;gt;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Maybe not easy understanding of so many option setup.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
We use option A.&lt;br /&gt;
OPA2111 set gain X 100.&lt;br /&gt;
AD8422 use as Gain X1 ( no need put R5 &amp;amp; R6 to AD8422 , because we use option A )&lt;br /&gt;
&lt;br /&gt;
How to get X100 :&lt;br /&gt;
R23 (100K) / R27 (2K) = 50&lt;br /&gt;
R24 (100K) / R27 (2K) = 50&lt;br /&gt;
50 + 50 = 100&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Then IC3D set gain X250&lt;br /&gt;
&lt;br /&gt;
How to get X250 :&lt;br /&gt;
1 + (R16 (1M) / R15 (4.02K) ) = 250 (249.75622)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Total : X 100 X 250 = X 25,000&lt;br /&gt;
&lt;br /&gt;
It is not X 10,000.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
C15 : 1 µF makes 0.16Hz low cut ( default )&lt;br /&gt;
C15 : 0.1 µF makes 1.6Hz low cut.&lt;br /&gt;
&lt;br /&gt;
C35 : 47 pF makes 3200 Hz high cut ( default )&lt;br /&gt;
C35 : 1 nF makes 160 Hz high cut&lt;br /&gt;
&lt;br /&gt;
Need setup for other channel too.&lt;br /&gt;
&lt;br /&gt;
[[File:Brain-duino_circuit.png]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;&amp;lt;Coming out of the first and second round of filters in IC3 on both channels we&#039;re seeing our signal clipping at +/-12V. Is this expected? &amp;gt;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From IC3 output, max clip is +-12V.&lt;br /&gt;
Usually not clip to +-12V. In case clip continue then not working.&lt;br /&gt;
Make sure solder jumper : SJ8 : 2-3 connect is default.&lt;br /&gt;
&lt;br /&gt;
C46 1µF is option.  default is put 0 ohm.&lt;br /&gt;
&lt;br /&gt;
==March 08 2017==&lt;br /&gt;
&lt;br /&gt;
Debugging the brainduino v0.1_foo ...&lt;br /&gt;
&lt;br /&gt;
One of these resistors is not like the other.  &lt;br /&gt;
&lt;br /&gt;
R15 is upside down ?&lt;br /&gt;
A 4.02 KOhm resistor.&lt;br /&gt;
(John wonders if this is a correct value for setting op-amp gain anyway,&lt;br /&gt;
given the corresponding 1 MOhm resistor in that part of the circuit.)&lt;br /&gt;
&lt;br /&gt;
R17 the same ... maybe not upside down afterall.  Jade observes that neither R15 or R17 have any visible printing on the side facing up, unlike any of the many many other resistors on the board.&lt;br /&gt;
Hence the initial speculation as to its potential upside-downness ...&lt;br /&gt;
&lt;br /&gt;
Breaking out the multimeter now ... Here we go ... What&#039;s going on with this multimeter ?&lt;br /&gt;
&lt;br /&gt;
Now reads 4.01 KOhm on R15&lt;br /&gt;
same on R17, very good.&lt;br /&gt;
&lt;br /&gt;
The feedback resistors R16 and R19:&lt;br /&gt;
R16: 1.001 MOhm&lt;br /&gt;
R19: 1.001 MOhm&lt;br /&gt;
&lt;br /&gt;
C35 difficult to get consistent reading of capacitance (maybe due to rest of circuit not isolated)&lt;br /&gt;
C36 11.85 nanoFarads ? Also not consistent .007 nanoFarads ... 0.4 ... 2. ... 1. ... 0.006&lt;br /&gt;
ok, we want 0.047 nanoFarads ie 47 picoFarads as specified in schematic&lt;br /&gt;
but we never see that ...&lt;br /&gt;
but that&#039;s okay, because measuring capacitance in circuit is a fool&#039;s errand.&lt;br /&gt;
&lt;br /&gt;
backing up a bit ...&lt;br /&gt;
R3 tests within 2% of 1 MegaOhm despite being GROSSLY OFF CENTER&lt;br /&gt;
R4 1.000 MegaOhm feels more solid than R3&lt;br /&gt;
&lt;br /&gt;
should we resolder R3 ?&lt;br /&gt;
&lt;br /&gt;
John would (if not feeling adventurous) read documentation&lt;br /&gt;
and study IC 1b and IC 1a ...&lt;br /&gt;
&lt;br /&gt;
Dan, feeling adventurous, reads Masahiro&#039;s instructables article&lt;br /&gt;
which is full of win, and yields A CLUE as per corner frequency of the&lt;br /&gt;
EIGHT POLE BUTTERWORTH FILTER: &amp;quot;1/100 of clock is filter high cut frequency.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==December 21 2016==&lt;br /&gt;
Understanding the brainduino v0.1&lt;br /&gt;
&lt;br /&gt;
Power supplies, voltage/current regulator. voltage reference&lt;br /&gt;
&lt;br /&gt;
LT1761 - 100mA, Low Noise, LDO Micropower Regulators &lt;br /&gt;
(IC7) (5 Volts for DC offset, 2.5 V after voltage divider,&lt;br /&gt;
feeds + signal inputs of 2 different op amps, also provides 5 V reference for Arduino)&lt;br /&gt;
&lt;br /&gt;
RB-0512 - +/- 12 V for op amp (RB-0512 D/P -- recom 12v dc/dc converter)&lt;br /&gt;
(drives all op amps, also feeds LT1761)&lt;br /&gt;
&lt;br /&gt;
Voltage divider (cf LT1761) built around SJ10, SJ11&lt;br /&gt;
a voltage divider for each channel ...&lt;br /&gt;
also linked with capacitors (two 10 microFarad on input, 0.1 microFarad for each output channel) &lt;br /&gt;
&lt;br /&gt;
Op Amps&lt;br /&gt;
* for amplification (instrumentation amp setup)&lt;br /&gt;
* for filtering (capacitors C15 and C11&lt;br /&gt;
&lt;br /&gt;
Butterworth filter&lt;br /&gt;
&lt;br /&gt;
Arduino (ADC, communication with Bluetooth module)&lt;br /&gt;
&lt;br /&gt;
... examining the hardware:&lt;br /&gt;
question about IC2 - funky solder, corrosion?&lt;br /&gt;
First amplifier of channel one (IC2A and IC2B)&lt;br /&gt;
&lt;br /&gt;
verified IC5, IC6 not populated ...&lt;br /&gt;
&lt;br /&gt;
still looking for IC7!&lt;br /&gt;
found it.  It has some funky solder right next to it.&lt;br /&gt;
IC7 is tiny.  Is it actually IC7?&lt;br /&gt;
Unlabeled solder blob next to it has white rectangle printed around it ... so maybe solder jumper&lt;br /&gt;
&lt;br /&gt;
IC7 connects to C32&lt;br /&gt;
&lt;br /&gt;
SJ16 is one step closer to bringing voltage divider into effect.&lt;br /&gt;
&lt;br /&gt;
Can measure on blob for 5 Volts ...&lt;br /&gt;
&lt;br /&gt;
There may be some test points near U$7&lt;br /&gt;
&lt;br /&gt;
Some resistors from divider network identified?&lt;br /&gt;
R39 is 4.3 K Ohms&lt;br /&gt;
R34 and R35 are each 10 K Ohms ?&lt;br /&gt;
both contact R29 which is 2.15 K Ohms to ground&lt;br /&gt;
&lt;br /&gt;
[[Image:IMG_9574.jpg|600px]]&lt;br /&gt;
[[Image:IMG_9573.jpg|600px]]&lt;br /&gt;
&lt;br /&gt;
==December 14 2016==&lt;br /&gt;
what is value of c33? 50 pF or 1 nF ?&lt;br /&gt;
determines high-cut frequency either 160 Hz (1 nF) or 3200 Hz (50 pF)&lt;br /&gt;
&lt;br /&gt;
Op amps ...&lt;br /&gt;
&lt;br /&gt;
OPA2111 - what gain? expecting ~ 100 ?&lt;br /&gt;
&lt;br /&gt;
what values for R27, R23, R24 ?&lt;br /&gt;
(also R28, R25, R26)&lt;br /&gt;
&lt;br /&gt;
R27 should be 2K ohms&lt;br /&gt;
R23, R24 -&amp;gt; 100K ohms&lt;br /&gt;
&lt;br /&gt;
cf instrumentation amp (figure 9 of http://www.ti.com/lit/ds/symlink/opa2111.pdf)&lt;br /&gt;
1 + (R23 + R24) / R27 ...&lt;br /&gt;
so 1 + (200 K / 2 K) = 101 &lt;br /&gt;
&lt;br /&gt;
so if input signal is on order of 100 microvolts&lt;br /&gt;
then output of first amp will be 10.100 millivolts &lt;br /&gt;
(really more like -50 to +50 microvolts -&amp;gt; -5 to +5 millivolts)&lt;br /&gt;
&lt;br /&gt;
AD8422&lt;br /&gt;
R5 and R6 are 200 ohms&lt;br /&gt;
G = 1 + (19.8 kΩ/RG)&lt;br /&gt;
= 1 + 99 = 100&lt;br /&gt;
&lt;br /&gt;
??? also see fig. 56 on page 20 of http://www.analog.com/media/en/technical-documentation/data-sheets/AD8422.pdf &lt;br /&gt;
&lt;br /&gt;
now consider if input to AD8422 is on order of 10 millivolts&lt;br /&gt;
output will be 1.000 volts  &lt;br /&gt;
(or if no offset voltage and input is actually ~ -5 to +5 millivolts,&lt;br /&gt;
then output will be -0.5 to +0.5 volts)&lt;br /&gt;
&lt;br /&gt;
(however, there is a +2.5 Volt DC offset impinging on the AD2177 part of the instrumentation amp setup&lt;br /&gt;
... we think this is ok to say ... But is option A grounded -- no offset -- or 2.5 volts ? &lt;br /&gt;
&lt;br /&gt;
where might DC offset voltage come into play? &lt;br /&gt;
&lt;br /&gt;
2nd Stage AD8422 Instrumentation amp with gain resistor&lt;br /&gt;
2(10 kΩ + RREF)/(20 kΩ + RREF)&lt;br /&gt;
http://www.analog.com/media/en/technical-documentation/data-sheets/AD8422.pdf&lt;br /&gt;
In figure 56, you will see that an op amp is used to input to the VREF pin. &lt;br /&gt;
&lt;br /&gt;
3rd stage Low-cut (aka high-pass ) filter and DC offset to 3rd Stage&lt;br /&gt;
OP2177ARMZ ( IC1B Ch1 and IC1A Ch2 )&lt;br /&gt;
&lt;br /&gt;
4th Stage IC3D&amp;gt;IC3C Ch1, and IC3A&amp;gt;IC3B, Potentially considered a two-pole high-cut(aka low-pass)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Next week: verify 101 x gain, 100 x gain, DC offset !&lt;/div&gt;</summary>
		<author><name>198.27.140.180</name></author>
	</entry>
	<entry>
		<id>https://wiki.extremist.software/index.php?title=Talk:DreamTeam/Brainduino&amp;diff=57400</id>
		<title>Talk:DreamTeam/Brainduino</title>
		<link rel="alternate" type="text/html" href="https://wiki.extremist.software/index.php?title=Talk:DreamTeam/Brainduino&amp;diff=57400"/>
		<updated>2017-03-22T03:06:12Z</updated>

		<summary type="html">&lt;p&gt;198.27.140.180: March 21 2017&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;==March 21 2017==&lt;br /&gt;
Q&amp;amp;A with Masahiro!&lt;br /&gt;
&amp;lt;&amp;lt;&lt;br /&gt;
We are making progress on our Brainduino V0.1 assembly. We think we isolated the abnormality to IC3, the low-pass filters. We observe 10,000x gain out of the first two rounds of amplification. Our understanding is that 10,000x gain amplifies a +/-50uV brain powered signal to a +/-0.5V signal, within the operational range of the Arduino ACD.&lt;br /&gt;
&amp;gt;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
Maybe not easy understanding of so many option setup.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
We use option A.&lt;br /&gt;
OPA2111 set gain X 100.&lt;br /&gt;
AD8422 use as Gain X1 ( no need put R5 &amp;amp; R6 to AD8422 , because we use option A )&lt;br /&gt;
&lt;br /&gt;
How to get X100 :&lt;br /&gt;
R23 (100K) / R27 (2K) = 50&lt;br /&gt;
R24 (100K) / R27 (2K) = 50&lt;br /&gt;
50 + 50 = 100&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Then IC3D set gain X250&lt;br /&gt;
&lt;br /&gt;
How to get X250 :&lt;br /&gt;
1 + (R16 (1M) / R15 (4.02K) ) = 250 (249.75622)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Total : X 100 X 250 = X 25,000&lt;br /&gt;
&lt;br /&gt;
It is not X 10,000.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
C15 : 1 µF makes 0.16Hz low cut ( default )&lt;br /&gt;
C15 : 0.1 µF makes 1.6Hz low cut.&lt;br /&gt;
&lt;br /&gt;
C35 : 47 pF makes 3200 Hz high cut ( default )&lt;br /&gt;
C35 : 1 nF makes 160 Hz high cut&lt;br /&gt;
&lt;br /&gt;
Need setup for other channel too.&lt;br /&gt;
&lt;br /&gt;
[[File:Brain-duino_circuit.png]]&lt;br /&gt;
&lt;br /&gt;
&amp;lt;&amp;lt;Coming out of the first and second round of filters in IC3 on both channels we&#039;re seeing our signal clipping at +/-12V. Is this expected? &amp;gt;&amp;gt;&lt;br /&gt;
&lt;br /&gt;
From IC3 output, max clip is +-12V.&lt;br /&gt;
Usually not clip to +-12V. In case clip continue then not working.&lt;br /&gt;
Make sure solder jumper : SJ8 : 2-3 connect is default.&lt;br /&gt;
&lt;br /&gt;
C46 1µF is option.  default is put 0 ohm.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==March 08 2017==&lt;br /&gt;
&lt;br /&gt;
Debugging the brainduino v0.1_foo ...&lt;br /&gt;
&lt;br /&gt;
One of these resistors is not like the other.  &lt;br /&gt;
&lt;br /&gt;
R15 is upside down ?&lt;br /&gt;
A 4.02 KOhm resistor.&lt;br /&gt;
(John wonders if this is a correct value for setting op-amp gain anyway,&lt;br /&gt;
given the corresponding 1 MOhm resistor in that part of the circuit.)&lt;br /&gt;
&lt;br /&gt;
R17 the same ... maybe not upside down afterall.  Jade observes that neither R15 or R17 have any visible printing on the side facing up, unlike any of the many many other resistors on the board.&lt;br /&gt;
Hence the initial speculation as to its potential upside-downness ...&lt;br /&gt;
&lt;br /&gt;
Breaking out the multimeter now ... Here we go ... What&#039;s going on with this multimeter ?&lt;br /&gt;
&lt;br /&gt;
Now reads 4.01 KOhm on R15&lt;br /&gt;
same on R17, very good.&lt;br /&gt;
&lt;br /&gt;
The feedback resistors R16 and R19:&lt;br /&gt;
R16: 1.001 MOhm&lt;br /&gt;
R19: 1.001 MOhm&lt;br /&gt;
&lt;br /&gt;
C35 difficult to get consistent reading of capacitance (maybe due to rest of circuit not isolated)&lt;br /&gt;
C36 11.85 nanoFarads ? Also not consistent .007 nanoFarads ... 0.4 ... 2. ... 1. ... 0.006&lt;br /&gt;
ok, we want 0.047 nanoFarads ie 47 picoFarads as specified in schematic&lt;br /&gt;
but we never see that ...&lt;br /&gt;
but that&#039;s okay, because measuring capacitance in circuit is a fool&#039;s errand.&lt;br /&gt;
&lt;br /&gt;
backing up a bit ...&lt;br /&gt;
R3 tests within 2% of 1 MegaOhm despite being GROSSLY OFF CENTER&lt;br /&gt;
R4 1.000 MegaOhm feels more solid than R3&lt;br /&gt;
&lt;br /&gt;
should we resolder R3 ?&lt;br /&gt;
&lt;br /&gt;
John would (if not feeling adventurous) read documentation&lt;br /&gt;
and study IC 1b and IC 1a ...&lt;br /&gt;
&lt;br /&gt;
Dan, feeling adventurous, reads Masahiro&#039;s instructables article&lt;br /&gt;
which is full of win, and yields A CLUE as per corner frequency of the&lt;br /&gt;
EIGHT POLE BUTTERWORTH FILTER: &amp;quot;1/100 of clock is filter high cut frequency.&amp;quot;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==December 21 2016==&lt;br /&gt;
Understanding the brainduino v0.1&lt;br /&gt;
&lt;br /&gt;
Power supplies, voltage/current regulator. voltage reference&lt;br /&gt;
&lt;br /&gt;
LT1761 - 100mA, Low Noise, LDO Micropower Regulators &lt;br /&gt;
(IC7) (5 Volts for DC offset, 2.5 V after voltage divider,&lt;br /&gt;
feeds + signal inputs of 2 different op amps, also provides 5 V reference for Arduino)&lt;br /&gt;
&lt;br /&gt;
RB-0512 - +/- 12 V for op amp (RB-0512 D/P -- recom 12v dc/dc converter)&lt;br /&gt;
(drives all op amps, also feeds LT1761)&lt;br /&gt;
&lt;br /&gt;
Voltage divider (cf LT1761) built around SJ10, SJ11&lt;br /&gt;
a voltage divider for each channel ...&lt;br /&gt;
also linked with capacitors (two 10 microFarad on input, 0.1 microFarad for each output channel) &lt;br /&gt;
&lt;br /&gt;
Op Amps&lt;br /&gt;
* for amplification (instrumentation amp setup)&lt;br /&gt;
* for filtering (capacitors C15 and C11&lt;br /&gt;
&lt;br /&gt;
Butterworth filter&lt;br /&gt;
&lt;br /&gt;
Arduino (ADC, communication with Bluetooth module)&lt;br /&gt;
&lt;br /&gt;
... examining the hardware:&lt;br /&gt;
question about IC2 - funky solder, corrosion?&lt;br /&gt;
First amplifier of channel one (IC2A and IC2B)&lt;br /&gt;
&lt;br /&gt;
verified IC5, IC6 not populated ...&lt;br /&gt;
&lt;br /&gt;
still looking for IC7!&lt;br /&gt;
found it.  It has some funky solder right next to it.&lt;br /&gt;
IC7 is tiny.  Is it actually IC7?&lt;br /&gt;
Unlabeled solder blob next to it has white rectangle printed around it ... so maybe solder jumper&lt;br /&gt;
&lt;br /&gt;
IC7 connects to C32&lt;br /&gt;
&lt;br /&gt;
SJ16 is one step closer to bringing voltage divider into effect.&lt;br /&gt;
&lt;br /&gt;
Can measure on blob for 5 Volts ...&lt;br /&gt;
&lt;br /&gt;
There may be some test points near U$7&lt;br /&gt;
&lt;br /&gt;
Some resistors from divider network identified?&lt;br /&gt;
R39 is 4.3 K Ohms&lt;br /&gt;
R34 and R35 are each 10 K Ohms ?&lt;br /&gt;
both contact R29 which is 2.15 K Ohms to ground&lt;br /&gt;
&lt;br /&gt;
[[Image:IMG_9574.jpg|600px]]&lt;br /&gt;
[[Image:IMG_9573.jpg|600px]]&lt;br /&gt;
&lt;br /&gt;
==December 14 2016==&lt;br /&gt;
what is value of c33? 50 pF or 1 nF ?&lt;br /&gt;
determines high-cut frequency either 160 Hz (1 nF) or 3200 Hz (50 pF)&lt;br /&gt;
&lt;br /&gt;
Op amps ...&lt;br /&gt;
&lt;br /&gt;
OPA2111 - what gain? expecting ~ 100 ?&lt;br /&gt;
&lt;br /&gt;
what values for R27, R23, R24 ?&lt;br /&gt;
(also R28, R25, R26)&lt;br /&gt;
&lt;br /&gt;
R27 should be 2K ohms&lt;br /&gt;
R23, R24 -&amp;gt; 100K ohms&lt;br /&gt;
&lt;br /&gt;
cf instrumentation amp (figure 9 of http://www.ti.com/lit/ds/symlink/opa2111.pdf)&lt;br /&gt;
1 + (R23 + R24) / R27 ...&lt;br /&gt;
so 1 + (200 K / 2 K) = 101 &lt;br /&gt;
&lt;br /&gt;
so if input signal is on order of 100 microvolts&lt;br /&gt;
then output of first amp will be 10.100 millivolts &lt;br /&gt;
(really more like -50 to +50 microvolts -&amp;gt; -5 to +5 millivolts)&lt;br /&gt;
&lt;br /&gt;
AD8422&lt;br /&gt;
R5 and R6 are 200 ohms&lt;br /&gt;
G = 1 + (19.8 kΩ/RG)&lt;br /&gt;
= 1 + 99 = 100&lt;br /&gt;
&lt;br /&gt;
??? also see fig. 56 on page 20 of http://www.analog.com/media/en/technical-documentation/data-sheets/AD8422.pdf &lt;br /&gt;
&lt;br /&gt;
now consider if input to AD8422 is on order of 10 millivolts&lt;br /&gt;
output will be 1.000 volts  &lt;br /&gt;
(or if no offset voltage and input is actually ~ -5 to +5 millivolts,&lt;br /&gt;
then output will be -0.5 to +0.5 volts)&lt;br /&gt;
&lt;br /&gt;
(however, there is a +2.5 Volt DC offset impinging on the AD2177 part of the instrumentation amp setup&lt;br /&gt;
... we think this is ok to say ... But is option A grounded -- no offset -- or 2.5 volts ? &lt;br /&gt;
&lt;br /&gt;
where might DC offset voltage come into play? &lt;br /&gt;
&lt;br /&gt;
2nd Stage AD8422 Instrumentation amp with gain resistor&lt;br /&gt;
2(10 kΩ + RREF)/(20 kΩ + RREF)&lt;br /&gt;
http://www.analog.com/media/en/technical-documentation/data-sheets/AD8422.pdf&lt;br /&gt;
In figure 56, you will see that an op amp is used to input to the VREF pin. &lt;br /&gt;
&lt;br /&gt;
3rd stage Low-cut (aka high-pass ) filter and DC offset to 3rd Stage&lt;br /&gt;
OP2177ARMZ ( IC1B Ch1 and IC1A Ch2 )&lt;br /&gt;
&lt;br /&gt;
4th Stage IC3D&amp;gt;IC3C Ch1, and IC3A&amp;gt;IC3B, Potentially considered a two-pole high-cut(aka low-pass)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Next week: verify 101 x gain, 100 x gain, DC offset !&lt;/div&gt;</summary>
		<author><name>198.27.140.180</name></author>
	</entry>
</feed>