Assorted Ideas - fcrimins/fcrimins.github.io GitHub Wiki
Patent: Semantic bookmark search
- Similar to gmail search, except rather than searching the contents of emails, you're searching the contents of Web pages.
- The problem with the semantic Web is that different observers have different opinions of what things mean
SellersPlace.com
DopaBlocker
3-Star Sentiment
Learn differences, not similarities
- Game Theory Reveals the Future of Deep Learning
- Real learning happens by learning differences, not similarities. Humans don't learn the difference between cats and dogs by looking at 1000 cats and 1000 dogs and building models of each. A human may see a cat one day, then a dog a few days later, then a horse a few days after that, and then a picture of a cat again. Yet somehow Cat, as a concept, is learned.
- This starts with an empty container suited for holding the representation of a concept. The empty concept container doesn't start out empty however, it gets initialized with the features of all other concepts negated--the "nots" of everything else. These "nots," these differences, are removed (think: marble sculpting) when it is learned that they do not apply. So, for example, when Cat is learned, the Cat Container starts out containing {not Dog, not Horse} as well as the negations of all of their accompanying characteristics.
- Perhaps this is related to Bayesian fitting (as opposed to Frequentist). An empty Concept Container starts out with every model and then learns which are more probable by whittling away those that don't fit the data well.
- The grid search is through feature space (with 1000 different features/models to start), not through concept instance space (with 1000 different instances of cats to start). I.e. search models, don't search data points.
- "specifies a communication protocol that tracks how often an algorithm makes queries about the objective" [https://arxiv.org/pdf/1509.08627.pdf] i.e. learn Cat from {2 cats a dog and a horse}, not from {1000 cats}
- update 3/22/17 - neural episodic control How DeepMind’s Memory Trick Helps AI Learn Faster “Our architecture does not try to learn when to write to memory, as this can be slow to learn and take a significant amount of time,” say Pritzel and co. “Instead, we elect to write all experiences to the memory, and allow it to grow very large compared to existing memory architectures.”
- FWC - this also makes me think that machine learning will need some notion of sleep to perform well--perhaps sleep is what weeds out the crap experiences that have been written to memory (as suggested by some article that I wish I could search my notes for right now--but I actually can!--I just grepped my wiki for "sleep" and saw the "Crick and Mitchison" reference)
- update 5/15/17 - Navigating the Unsupervised Learning Landscape
- "Ideally, we would like to have a model that behaves more like our brain. That needs just a few labels here"
- "the most successful models are the ones that predict future representation of a video"
- encoding: "the primary visual vortex (V1) in our brain uses principles of sparsity to create a minimal set of base functions that can be also used to reconstruct the input image."
- "Auto-encoders / sparse-coding / stacked auto-encoders advantages and disadvantages
- Advantages:
- simple technique: reconstruct input
- multiple layer can be stacked
- intuitive and based on neuroscience research
- Disadvantages:
- each layer is trained greedily
- no global optimization
- does not match performance of supervised learning
- multiple layer ineffective
- reconstruction of input may not be ideal metric for learning a general-purpose representation"
- Advantages:
- "Generative models try to create a categorization (discriminator or encoder) network and a model that generates images (generative model) at the same time."
- "Generative adversarial model advantages and disadvantages
- Advantages:
- global training of entire network
- simple to code and implement
- Disadvantages:
- hard to train, conversion problems
- in some cases matches performance of supervised learning
- need to prove usability of representation (a problem of ALL unsupervised algorithms)"
- Advantages:
- Unsupervised "Learn-from-data models"
- Clever tricks:
- "break the image into a puzzle and train a deep neural network to solve the puzzle"
- two patches from same image + another patch from third -> train discriminator
- Clever tricks:
- "Unsupervised training is very much an open topic, where you can make a large contribution by:
- creating a new unsupervised task to train networks, e.g.: solve a puzzle, compare image patches, generate images, …)
- thinking of tasks that create great unsupervised features, e.g.: what is object and what is background, same on stereo images, same on video frames ~= similar to how our human visual system develops"
Selective HTTP (12/23/16)
- Once a page has been served to a client, the thing doing the serving (e.g. the browser) should know what the important components of the page are (e.g. by what users choose to highlight, or ?) such that future requests can be for only the important components.
- This may be useful for low bandwidth connections: How to Share Your Phone’s Internet Connection
Record my life; supplement my brain
- everything I read, I may want to go back to it and search (e.g. in the fall of last year i had a conversation with someone about a band)
- The Great AI Awakening (at Google) (12/17/16)
- "One is not what he is for what he writes, but for what he has read."
- "If you wanted to translate from English to Japanese, for example, you would program into the computer all of the grammatical rules of English, and then the entirety of definitions contained in the Oxford English Dictionary, and then all of the grammatical rules of Japanese, as well as all of the words in the Japanese dictionary, and only after all of that feed it a sentence in a source language and ask it to tabulate a corresponding sentence in the target language. You would give the machine a language map that was, as Borges would have had it, the size of the territory. This perspective is usually called 'symbolic A.I.' — because its definition of cognition is based on symbolic logic — or, disparagingly, 'good old-fashioned A.I.'"
- FWC - so the question is then, can "loose" symbolic rules, like in language, be efficiently incorporated into learned systems to make them better? perhaps simply by inputting them as features and then letting the system "learn" how helpful they are. this would perhaps take pressure off the rest of the network. but would we then want to tell the network that these inputs are "different" and not to get rid of them as easily as others perhaps?
- "The neuronal 'voters' will recognize a happy cat dozing in the sun ... as long as they have been exposed to millions of diverse cat scenes."
- FWC - but then why don't humans need millions of such exposures to cats to understand what a cat looks like? => there is still something fundamentally different between current AI and human I. perhaps it has to do with all the pictures of things humans see that aren't cats. humans see millions of things, they're just of all sorts of different things, and they learn negation, "not a cat" .... or maybe humans do see millions of pictures of cats, they're just all streamed together over time, like frames in a video.... this suggests that there should be a time dimension to hidden NN layers
- FWC - can an algorithm fed with video of a single cat learn to identify other cats? this algorithm would be closer to the human brain -- many different dimensions of a single item lead to understanding of what makes that item what it is
- "Another was that he wasn’t at all embarrassed to say sincere things like 'if we put our minds to it.'"
- "The benchmark metric to evaluate machine translation is called a BLEU score"
- "The team had in their storehouse about 97 million unique English 'words.' But once they removed the emoticons, and the misspellings, and the redundancies, they had a working vocabulary of only around 160,000."
- FWC - this article makes me want to implement a summarizer - remove the most irrelevant or duplicated piece of this article --one sentence at a time (similar to my summarizer idea for redundant (reality) tv shows) -- similar to my idea to "dedup the web" - though different readers might have different opinions of which sentences are important!
There are two main problems with the old-fashioned approach. The first is that it’s awfully time-consuming on the human end. The second is that it only really works in domains where rules and definitions are very clear: in mathematics, for example, or chess
Train a NN to predict a word based on its definition (from a dictionary)
- or use such a mapping with negative sampling, to construct word embeddings
- aren't such things already trained on Wikipedia though?
Machine Learning of Neural Net Architecture
- Eg detect the effectiveness of individual nodes, and allow them to die off or multiply based on how effective they are.
- Allow tumors to form similar to how LSTMs have cell state plus regular CNN. How might a CNN detect that it needs a cell state for example?
- This all still requires a predefined cost function. ie, what to learn. How could that be learned? seems like theres a finite set of these tho, squared error, cross entropy, logisti
- This idea seems similar to dropout, but deletion is only one of the forms of mutation, and plus its mainly done to prevent overfitting, not to learn the architecture of the net
- Learning to learn by gradient descent by gradient descent (in TensorFlow)
Linguistic Networks
Chem Collection (8/17/16)
- There are tons of fertilizers flowing into the Gulf of Mexico out the mouth of the Mississippi River.
- Harvest them to (1) remove the chemicals and (2) purify the Gulf.
My reality show idea (4/27/16)
- Talkshow is texting in public: http://kottke.org/16/04/talkshow-is-texting-in-public
A programming language...
- ...where every function can think for itself, decide when it wants to run based on patterns of use
- every menu option can show itself when it wants (e.g. if the user navigates to it many times, then it can make itself easier to navigate to)
- e.g. I so often send email messages w/out a body, and I don't want my phone to ask me if that's really what I want to do every time
All software should adapt
- this means building on top of an adaptive platform
- i go to 3 options in every eclipse pop-up menu
- i only "share" from my phone to gmail and text messaging
- when i type "joseph" i typically only mean one person
- expected-next-word language modeling is where this might work correctly, but this should be a feature of something lower down than nlp that should percolate up to all these other places also
How the Internet Became Commercial - Counter argument to government invented the internet (4/15/16)
Elasticsearch as a Time Series Database - Getting data in with StatsD (2/20/16)
The UNIX School: awk & sed tutorial (1/25/16)
Getting Started with Git (Channel 9) (1/25/16)
Rust and the Blub Paradox (1/28/16)
Manager Strategies: People can read their manager's mind (1/19/16)
Finely-Grained Management (2/7/16)
- Heisenberg Developers: You can not observe a developer without altering their behavior.
Nice git cheatsheet (12/7/15)
Overconfidence (1/2/16)
- RCTs for public policy. Why is introspection all that is currently required? People accept that the human body is a complex system, requiring RCTs, but they don't accept that an economy is.
Why GNU grep is fast (12/10/15)
Why use REST inside the company? (12/9/15)
Email: Use anomaly detection for SIDS (10/6/15)
- SIDS is an anomaly detection scenario because there are very few cases (e.g. 0.1%).
Email: Don't regulate pay (10/6/15)
- It's been tried and it doesn't work. Pay in finance just keeps going up.
- Pay that's earned should be taxed less than pay that isn't. Another way to think about this is pay that's acquired by chance should be taxed more than pay that isn't.
- So then we just need a metric of chance for each industry/sector/other set of factors. One idea: the ratio of successful to unsuccessful members of an industry.
The Warmth of Other Suns (10/10/15)
Book:- "The Warmth of Other Suns is about the Great Migration, the mass movement of African Americans from the Southern US to the Northeast, Midwest, and West between 1910 and 1970. During that time, roughly 6 million African Americans moved north and west to escape Jim Crow laws, discrimination, low wages, the threat of physical violence & death, and everyday humiliation & lack of freedom in the South."
- Sound like anything going on today, e.g. Mexicans perhaps. So why isn't it being celebrated like it is in this book?
- What's more, hasn't this been a constant in the history of America?
Charter Prisons (Email: "Why don't", 10/17/15)
- There are private prisons but are there non profit prisons? Charter prisons?
- Wouldn't it be possible to massively increase prisoner welfare by just attaching GPS and/or listening/recording devices to their ankles that would set off alarms if tampered with.
- This could protect them against other prisoners and guards.
- Could even make it voluntary, if prisoners objected on privacy grounds--not that there's any of that in prison--not that I know.
Facial Biases (10/20/15)
- Email: Thin upper lip bias
- It seems like folks with thin upper lips are assumed to be more dishonest or shifty than others. I had this idea while watching episode 39 of House of Cards.
- What other kinds of facial feature biases are there? Seems easy enough to learn. Would mean understanding features more than lines though.
- There are obviously race and sex biases. And there are biases against age and height. There are surely biases against all sorts of things we aren't aware of.
2 Problems with Maps Apps
- They aren't predictive. E.g. what will the traffic be like on a summer Friday afternoon leaving the city. I.e. they aren't forward looking. But they're not backwards looking either. If there's unexpected traffic (e.g. due to an accident) and all of a sudden traffic begins to flow freely again, that should be accounted for very quickly.
- The graph being traversed shouldn't have intersections as its nodes; rather nodes should be points before and after intersections because it might take longer to turn left than go straight. Take for example, exiting off of Rte 91 North onto Rte 84 East.
Email: The only way to learn is to mess up (10/1/15)
- Even theory-based learning requires the testing of theories
- i.e. using controls where on one side of the "line of control" lies the theory and the other side is "messed up" (i.e. the anti-theory)
9/25/15
DATA!
- IDEA: http://www.data.gov/
Email: Create a company...
- ...that rather than trying to always make things better, like google, tries to prevent things from becoming worse, by stifling innovation for example. Every cost against rent seeking
Email: ML for aggregated knowledge
- On the Web for example. How many sites support a particular point of view
- Internetal Consensus!
Idea: Feature Engineering for input and output features
- Not only do we want systems to learn the input features automatically, we should also want them to learn the output features automatically.
- We want them to learn what types of questions might be asked. They should act as a sort of database query language.
- E.g. given data.gov, we should be able to ask questions about any entity (the economy, specific companies or countries) that might have a relationship to that sort of data
Email: data support via ML for spoken/written words (9/17/15)
- People say all sorts of shit, but its very rarely backed up by data. It would be really cool if given some statement, one could use ML to look for support for that statement in all the data.
Email: back data out of published results (9/17/15)
- just like in the ML course where features are backed out of ratings (low rank matrix factorization)
- if data could be backed out of published results, then the data could be combined with data from other studies and "meta-data studies" could be born
Email: ML for myths
- Eg diets
- Things with tons of unproven "knowledge"
- Can myths/diets be debunked using ML?
Email: ML for understanding babies
- How many theories are there? How many don't have any supporting evidence--like psychology?
- Understand what their cries and motions mean
Email: Babies whack themselves in their faces (10/1/15)
- Because they're learning, just like back propagation.
Email: Newest cyber threat will be data manipulation, US intelligence chief says | Technology | The Guardian
- ML cyber security. All data must be publicly confirmable--like with Bitcoin.
- http://www.theguardian.com/technology/2015/sep/10/cyber-threat-data-manipulation-us-intelligence-chief
Email: Did the Chinese stock market crash BECAUSE US unemployment has dropped so low? Ie risk of rising rates? (9/17/15)
- I.e. contrary to what the following post suggests
- Why the Federal Reserve Decided to Wait a Few Months Before Messing With the Economy