EFTA02585481
EFTA02585482 DataSet-11
EFTA02585484

EFTA02585482.pdf

DataSet-11 2 pages 490 words document
P17 P24 V16 V11 P19
Open PDF directly ↗ View extracted text
👁 1 💬 0
📄 Extracted Text (490 words)
From: Joscha Bach Sent: Wednesday, October 11, 2017 5:56 PM To: Jeffrey Epstein Subject: Re: After skimming their paper, the idea seemed unexciting to me at first: =asically, if we have enough feature dimensions we can almost always =ind a linear separation. This is also related to how Support Vector =achines work: they project the data into an extremely high-dimensional =pace, find a separating hyperplane with linear regression, and then =roject that plane back into the original space as the separator. A =imilar idea is behind Echo State networks, which use a randomly wired =ecurrent neural network and then only train the output layer with a =ingle linear regression. The authors take an existing trained neural network, and whenever it =akes a mistake, they train a linear classifier on the network state and =ata, i.e. they try to find out when the network goes wrong. Instead of =mproving the network (which is also likely to make it worse in other =ases), they add an additional layer to it. For engineering, this makes = lot of sense, because large neural networks are cheap to use and =eploy but expensive to train. On a more philosophical level, it is tempting to ask if that might be a =eneral learning principle for brains: when you don't perform well, add =ore control structure on top. It probably makes sense whenever you are =onfident that training the existing structure won't improve it that =uch, but unless training the weights in an existing network, it also =dds quite a few milliseconds to the processing time. There is probably =n optimal tradeoff for this. The other thing is that the new layer is a =inear classifier only (at least in this paper), and it is creating a =ocal override on the system's results, instead of integrating with it, =omewhat similar to how reasoning might override our subconscious =ehavior. One of the drawbacks is that this won't allow us to use the =ew layer for simulating/understanding the structure of the domain =odeled by the rest of the network. — Joscha > On Oct 10, 2017, at 09:43, jeffrey E. <[email protected]> wrote: > https://www.sciencedaily.com/releases/2017/08/170821102725.htm > -- > please note > The information contained in this communication is confidential, may > be attorney-client privileged, may constitute inside information, and > is intended only for the use of the addressee. It is the property of > JEE Unauthorized use, disclosure or copying of this communication or > any part thereof is strictly prohibited and may be unlawful. If you > have received this communication in error, please notify us > immediately by return e-mail or by e-mail to [email protected], > and destroy this communication and all copies thereof, including all > attachments. copyright -all rights reserved <?xml version=.0" encoding=TF-8"?> <IDOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/Propertylist-1.0.dtd"> <plist version=.0"> alict> <key>conversation-idgkey> EFTA_R1_01765067 EFTA02585482 <Integer>28259</integer> <key>date-last-viewed</key> <integer>0</integer> <key>date-received</key> <integer>1507744549</integer> <key>flags</key> <integer>8590195717</integer> <key>gmail-label-ids</key> <array> <integer>6</integer> <integer>2</integer> </array> <key>remote-id</key> <string>757802</string> </diet> </plist> 2 EFTA_R1_01765068 EFTA02585483
ℹ️ Document Details
SHA-256
690f83af118587ac98dbeb6860660acc888b63b3ef45bf5708407128251d1823
Bates Number
EFTA02585482
Dataset
DataSet-11
Document Type
document
Pages
2

Comments 0

Loading comments…
Link copied!