EFTA02451626.pdf
PDF Source (No Download)
Extracted Text (OCR)
From:
on behalf of Ben Goertzel
Sent:
Monday, September 5, 2016 4:30 PM
To:
jeffrey E.
Subject:
Re: Emergence of Chomskyan "deep syntactic structure" via probabilistic inference
To pose it in a question form, hmmm....
I guess one could say
"It seems that, using a probabilistic-logic-based learning system, one obtains a system that LEARNS the deep syntactic
structure of a sentence, as an intermediate result DURING THE PROCESS OF learning surface syntactic structure based
on comparison of sentences with
their non-linguistic referents. This is because deep syntactic
structures often share more symmetries and patterns with the logical structure of sentences' non-linguistic referents
So a question would be: How could one tell, from the mere presence of deep syntactic structures as patterns in
language, whether (or the extent to which) such structures are innate in the brain versus created in the course of
learning (created due to their natural structural role as intermediaries between deep semantic structure and surface
syntactic structure)?"
Well that's kind of complicated, but these are not trivial matters I guess..
The case in point is that if one wants a system to learn the surface structure
"Who did Ben tickle?"
based on a nonlinguistic referent that has logical structure
tickle(Ben, ?)
then the probabilistic logic system will, in the course of learning, automatically construct
Ben did tickle who?
as part of its surface-syntax-learning process..
-- Ben
Regarding,
> i think he might ask , what happens with ben is tickling sue who is
> being ticked by bob while he is being tickled by both
EFTA_R1_01551213
EFTA02451626
-- of course that's a more complex case that requires more inference steps to handle, but since my approach is logic-
based, it doesn't have any problem with recursive constructs.... Recursive constructs are a problem for typical statistical
methods, which don't involve general abstract knowledge representation (just markov chains; or hierarchical patterns
with a fixed depth as in a typical deep neural net).... But probabilistic logic incorporates the "frequency counting" aspect
of statistical methods, AND the general abstract knowledge representation of logical methods (so it can represent
languages all the way up the good old "Chomsky formal language hierarchy"...) <?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/Propertylist-1.0.dtd">
<plist version="1.0">
<dict>
<key>conversation-id</key>
<integer>70236</integer>
<key>date-last-viewed</key>
<integer>0</integer>
<key>date-received</key>
<integer>1473093013</integer>
<key>flags</key>
<integer>8690195713</integer>
<key>gmail-label-ids</key>
<array>
<integer>6</integer>
<integer>2</integer>
</array>
<key>remote-id</key>
<string>641860</string>
</dict>
</plist>
2
EFTA_R1_01551214
EFTA02451627
Document Preview
PDF source document
This document was extracted from a PDF. No image preview is available. The OCR text is shown on the left.
This document was extracted from a PDF. No image preview is available. The OCR text is shown on the left.
Document Details
| Filename | EFTA02451626.pdf |
| File Size | 133.3 KB |
| OCR Confidence | 85.0% |
| Has Readable Text | Yes |
| Text Length | 2,933 characters |
| Indexed | 2026-02-12T17:18:39.222690 |