US12254265B2 - Generating unique word embeddings for jargon-specific tabular data for neural network training and usage - Google Patents
Generating unique word embeddings for jargon-specific tabular data for neural network training and usage Download PDFInfo
- Publication number
- US12254265B2 US12254265B2 US17/483,989 US202117483989A US12254265B2 US 12254265 B2 US12254265 B2 US 12254265B2 US 202117483989 A US202117483989 A US 202117483989A US 12254265 B2 US12254265 B2 US 12254265B2
- Authority
- US
- United States
- Prior art keywords
- entries
- cell
- context
- embeddings
- tabular data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 38
- 238000012549 training Methods 0.000 title claims description 16
- 238000000034 method Methods 0.000 claims abstract description 70
- 230000008569 process Effects 0.000 claims abstract description 21
- 210000004027 cell Anatomy 0.000 claims description 115
- 239000013598 vector Substances 0.000 claims description 34
- 230000015654 memory Effects 0.000 claims description 16
- 230000000873 masking effect Effects 0.000 claims description 10
- 210000004460 N cell Anatomy 0.000 claims description 6
- 230000004044 response Effects 0.000 claims description 2
- 238000010586 diagram Methods 0.000 description 16
- 230000006870 function Effects 0.000 description 16
- 238000004590 computer program Methods 0.000 description 9
- 238000012545 processing Methods 0.000 description 7
- 238000013459 approach Methods 0.000 description 5
- 238000004458 analytical method Methods 0.000 description 4
- 230000005540 biological transmission Effects 0.000 description 4
- 238000003058 natural language processing Methods 0.000 description 4
- 230000003287 optical effect Effects 0.000 description 3
- 230000004913 activation Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 2
- 230000001413 cellular effect Effects 0.000 description 2
- 238000013135 deep learning Methods 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 230000001537 neural effect Effects 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 241000699670 Mus sp. Species 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
- G06F40/177—Editing, e.g. inserting or deleting of tables; using ruled lines
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
- G06F40/177—Editing, e.g. inserting or deleting of tables; using ruled lines
- G06F40/18—Editing, e.g. inserting or deleting of tables; using ruled lines of spreadsheets
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
- G06F40/284—Lexical analysis, e.g. tokenisation or collocates
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
Definitions
- This invention generally relates to analysis of tabular information and, more specifically, relates to generating unique word embeddings for jargon-specific tabular data for neural network training and usage.
- tabular data can contain columns full of domain-specific jargon such as the following: alphanumeric codes; undeclared abbreviations; and the like.
- annotation of this data is generally very sparse, and the same entity could be referred to by multiple column titles. Further, a large number of tables may be distributed over a large number of systems.
- An exemplary method is disclosed of using a computing device to generate unique word embeddings for jargon-specific tabular data.
- the method includes accessing by a computing device tabular data containing a plurality of entries of alphanumeric data.
- the method includes generating, by the computing device using a tokenization process, a plurality of tokens of the plurality of entries of alphanumeric data.
- the tokenization process maintains jargon-specific features of the alphanumeric data.
- the method includes generating, by the computing device using the tokens, a plurality of embeddings of the plurality of entries of alphanumeric data.
- the embeddings capture similarity of the plurality of entries considering all of global features, column features, and row features in the tokens of the tabular data.
- the method includes predicting, by the computing device with a neural network, probabilities for pre-defined classes for the tabular data using the generated embeddings.
- a computing device that is used to generate unique word embeddings for jargon-specific tabular data.
- the computing device comprises one or more memories having computer-readable code thereon, and one or more processors/.
- the one or more processors in response to retrieval and execution of the computer-readable code, cause the computing device to perform operations comprising: accessing by the computing device tabular data containing a plurality of entries of alphanumeric data; generating, by the computing device using a tokenization process, a plurality of tokens of the plurality of entries of alphanumeric data, the tokenization process maintaining jargon-specific features of the alphanumeric data; generating, by the computing device using the tokens, a plurality of embeddings of the plurality of entries of alphanumeric data, the embeddings capturing similarity of the plurality of entries considering all of global features, column features, and row features in the tokens of the tabular data; and predicting, by the computing device with a neural network, probabilities for pre-
- a computer program product comprises a computer readable storage medium having program instructions embodied therewith.
- the program instructions are executable by a computing device to cause the computing device to perform operations comprising: accessing by the computing device tabular data containing a plurality of entries of alphanumeric data; generating, by the computing device using a tokenization process, a plurality of tokens of the plurality of entries of alphanumeric data, the tokenization process maintaining jargon-specific features of the alphanumeric data; generating, by the computing device using the tokens, a plurality of embeddings of the plurality of entries of alphanumeric data, the embeddings capturing similarity of the plurality of entries considering all of global features, column features, and row features in the tokens of the tabular data; and predicting, by the computing device with a neural network, probabilities for pre-defined classes for the tabular data using the generated embeddings.
- FIG. 1 is a block diagram of a framework for generating unique word embeddings for jargon-specific tabular data for neural network training and output in an exemplary embodiment
- FIG. 2 is block diagram of an exemplary and non-limiting system for implementation of the framework of FIG. 1 ;
- FIG. 3 which is spread over FIGS. 3 A and 3 B , is a logic flow diagram for generating unique word embeddings for jargon-specific tabular data for neural network training and usage in accordance with an exemplary embodiment
- FIG. 4 is a table illustrating results of tokenization for three different approached including an approach used herein in an exemplary embodiment.
- tables are rich in data and can provide vital information about an object. Extracting useful insights from tabular data may require domain expertise, however, especially if the information is composed of domain-specific jargon or codes.
- Tabular data may contain numerical data, categorical data, cohesive phrases with semantics, coded information, and the like, or a mix of these.
- Existing analysis methods rely on crucial assumptions regarding the type of data that the table may contain. These methods may be broadly categorized into two groups, depending on the type of data to which the method is applicable. There are methods that assume the data to be complete English phrases.
- the data can be assumed to be numerical or categorical. See Sercan O Arik and Tomas Pfister, “Tabnet: Attentive interpretable tabular learning”, arXiv preprint arXiv:1908.07442, 2019. In the Arik et al. paper, the authors use raw numerical values as direct input, and generate embeddings for categorical data. However, the authors do not handle non-categorical tabular data.
- Cell-Masking A tokenization method that generalizes the cell entries in tabular data resulting in better quality embeddings and compressed vocabulary.
- Cell2Vec A method to generate cell embedding by exploiting the row and column context.
- TableNN A supervised attention-based neural network that predicts cell category while being column-order invariant.
- This information regards tokenization, embedding, and attention-based neural networks.
- tokenization With respect to tokenization, this is an important and at times mandatory task while working with text data. Simply put, tokenization is breaking down the text into smaller chunks of text or tokens. These tokens make up the vocabulary for the model. Thus, the method used to generate these tokens is deemed important. In natural language applications, commonly used methods, such as word tokenization based on certain delimiters, character tokenization and subword tokenization, can provide reasonable results. However, they usually fail to encapsulate semantics in the way the text is tokenized.
- BPE byte pair encoding
- NLP Natural Language Processing
- a word is represented by a vector.
- BOW bag of word
- TF-IDF term frequency-inverse document frequency
- each vector entry is mapped to a word in the vocabulary, so that if a document contains that word, the related entry receives a non-zero value.
- the critical drawback for such representation is that the generated sparse vectors do not capture semantic similarity among the words.
- Word Embedding instead learns a dense vector from the surrounding context of each word, allows words with semantically similar meaning to have similar representations.
- For word embedding see Tomas Mikolov, Kai Chen, Gregory S. Corrado, and Jeffrey Dean, “Efficient estimation of word representations in vector space”, CoRR, abs/1301.3781, 2013.
- the Jiaoyan Chen et al. paper considers locality features that are extracted by application of convolution on the rows and columns surrounding the target cell.
- existing methods are useful when the cell entries have natural language meaning, they are not applicable when the cell entries have no obvious semantic meaning, and instead are just some form of domain-specific codes or jargon.
- the instant exemplary embodiments use jargon for representations that might be specific for an organization, and perhaps not for the whole trade group/profession.
- An example is the following.
- a tool in manufacturing facility may be referred to as CVD005, but that will be specific to the particular company, and perhaps the ‘CVD’ phrase has some meaning for the company.
- CVD005 A tool in manufacturing facility
- CVD005 a tool in manufacturing facility
- CVD005 a tool in manufacturing facility
- FIG. 1 this figure is a block diagram of a framework 100 for generating unique word embeddings for jargon-specific tabular data for neural network training and output in an exemplary embodiment.
- a table 110 has columns of date, unit ID (identification), and tool ID. Only two rows of the table 110 are illustrated.
- the cell-masking module 120 operates on the table 110 to create table 130 .
- the cell-masking module 120 is a tokenization method that generalizes the cell entries in tabular data. Tokenization is a process of breaking sentences into smaller pieces called tokens. Tokens are roughly equivalent to words, but not always, as described below. In this example, the numbers from the table 110 have been removed and replaced with “X” in the table 130 .
- the cell2vec module 140 uses the table 130 and generates cell embedding by exploiting the row and column context to create table 150 .
- Each NLP (natural language processing) token is translated into a digital representation (e.g., a vector) of the word.
- a digital representation e.g., a vector
- the cell and context embedding 155 are used by the tableNN (where NN is neural network) module 160 to create a trained NN model 170 .
- the tableNN module 160 may be a supervised attention-based neural network (as one example of a suitable NN) that predicts cell category while being column-order invariant.
- the trained NN is output via the output model 170 block.
- FIG. 2 this figure is a block diagram of an exemplary and non-limiting system 200 for implementation of the framework of FIG. 1 .
- the system 200 includes a computer system 210 , one or more wired or wireless networks 297 , and one or more other computer systems 290 .
- the computer system 210 is a computing device suitable for performing the exemplary embodiments herein.
- the computer system 210 includes one or more processors 220 , one or more memories 225 , one or more transceivers 230 , one or more network (N/W) interfaces (I/F(s)), user interface (I/F) circuitry 265 , and one or more antennas 228 .
- the user interface circuitry 265 may interface with one or more user interface elements 205 .
- the one or more memories 225 include computer program code 223 .
- the computer system 210 includes a control module 240 , comprising one of or both parts 240 - 1 and/or 240 - 2 , which may be implemented in a number of ways.
- the control module 240 may be implemented in hardware as control module 240 - 1 , such as being implemented as part of the one or more processors 220 .
- the control module 240 - 1 may be implemented also as an integrated circuit or through other hardware such as a programmable gate array.
- the control module 240 may be implemented as control module 240 - 2 , which is implemented as computer program code 223 and is executed by the one or more processors 120 .
- the one or more memories 225 and the computer program code 223 may be configured to, with the one or more processors 220 , cause the computer system 210 to perform one or more of the operations as described herein.
- the control module 240 implements at least the cell-masking module 120 , the cell2vec module 140 , and the tableNN module 160 and can implement the framework 100 of FIG. 1 .
- the one or more transceivers 230 include one or more wireless receivers (Rx) 232 and one or more wireless transmitters (Tx) 233 .
- the one or more transceivers 230 could be Bluetooth, NFC (near-field communication), Wi-Fi, satellite, cellular, or the like. These may interface with a wireless network 297 , such as a Wi-Fi and/or cellular and/or satellite network, via one or more wireless links 278 .
- the N/W I/F(s) 245 are wired interfaces that interface with a wired network 297 , via one or more wired links 277 .
- a user e.g., a human being
- the computer system 210 interfaces with the computer system 210 via one or more of the user I/F elements 205 , which can include camera(s), audio device(s) (such as speakers and/or microphones), display(s), input device(s) (such as mice or trackballs), and/or keyboards.
- the user I/F elements 205 could interface with the computer system 210 via the user I/F circuitry 265 , such as via a USB (universal serial bus) or via other circuitry.
- the user I/F elements 205 could interface with the computer system 210 via the transceivers 230 such as via Bluetooth.
- a user 291 uses one or more other computer systems 290 and interfaces with the computer system 210 and the control module 240 (and framework 100 ) via the wired or wireless networks 297 .
- the computer system 210 could be on the Internet, in a LAN (local area network) or part of the cloud, for instance.
- the computer readable memories 225 may be of any type suitable to the local technical environment and may be implemented using any suitable data storage technology, such as semiconductor-based memory devices, flash memory, firmware, magnetic memory devices and systems, optical memory devices and systems, fixed memory and removable memory.
- the computer readable memories 225 may be also means for performing storage functions.
- the processors 220 may be of any type suitable to the local technical environment, and may include one or more of general-purpose computers, special purpose computers, microprocessors, digital signal processors (DSPs) and processors based on a multi-core processor architecture, as non-limiting examples.
- the processors 220 may be means for performing functions, such as controlling the computer system 210 , and other functions as described herein.
- FIG. 3 is spread over FIGS. 3 A and 3 B , and which is a logic flow diagram for generating unique word embeddings for jargon-specific tabular data for neural network training and output.
- FIG. 3 illustrates the operation of an exemplary method or methods, a result of execution of computer program instructions embodied on a computer readable memory, functions performed by logic implemented in hardware, and/or interconnected means for performing functions in accordance with exemplary embodiments.
- the blocks in FIG. 3 are assumed to be performed by the computer system 210 , under control of the control module 240 , which implements the framework 100 of FIG. 1 .
- cell entries are processed by masking the numerical characters in the string, e.g., ABN29D is replaced with ABN**D, with the numerical part stored separately for later use for numerical frequency. See block 310 of FIG. 3 , where the computer system 210 processes cell entries in a table by masking numerical characters in strings and by storing the numerical characters.
- Such an approach is considered appropriate for tabular data because often times, the actual numbers are important enough to not to ignore them completely, but the minor variations between them may not be very important.
- NBA_2 K12 and NBA_2 K13 may be tokenized as NBA_*K**.
- FIG. 4 shows the results of tokenization using three difference approaches. That is, FIG. 4 is a table illustrating results of tokenization for three different approached including an approach used herein in an exemplary embodiment.
- the first row is text (corpus), indicating the starting point.
- This row is a sample row from a model attendance record of a school with the following cell entries: [Student ID, PRESENT/ABSENT(absentees), STUDENT NAME].
- the student ID is ABN29D, the student was absent twice (ABSENT2), and the student's name is ABHINAV. Tokens are separated by commas (,).
- the second row shows the tokens for the alpha-numeric tokenization method.
- the third row shows the tokens for the BPE (byte pair encoding) tokenization method.
- the last row shows the tokens for the exemplary embodiment of the cell-masking tokenization, where an asterisk (*) is used to replace the numbers, and the numbers that have been replaced are also shown.
- the numbers that have been replaced are kept and used for a numerical frequency, as described below.
- the numbers may also be used, to an extent, as differentiators (e.g., 29 is different from 25 ).
- the cell-masking tokenization process maintains the jargon-specific features of the input data. That is, the ABN..D of the text is assumed to have some meaning to the school that created the information, and does, as the information corresponds to the student ID. This language is therefore considered to be jargon and is specific to this school.
- the jargons-specific features illustrated here can be applied to any trade, profession, manufacturer or similar group, and can be processed to maintain these jargon-specific features via the exemplary cell-masking tokenization process disclosed herein.
- Context is selected in order to exploit the tabular structure, and can be considered to be the surrounding words for a target word. For instance, row context may consider all the cell entries in the row of the target cell. Column context may consider the next N column entries from the column of the target cell.
- row context may consider all the cell entries in the row of the target cell.
- Column context may consider the next N column entries from the column of the target cell.
- the features of Cell2Vec include consideration of all the relevant context (unlike word2vec), and this technique is column-order invariant.
- Cell2Vec considers the table as a document, each row as a sentence and masked strings as words (see block 315 of FIG. 3 ), the Cell2Vec model (as part of module 130 ) can be trained with a vector size of, e.g., 54 bytes, though this vector size is merely one example. Thus, this exploits the tabular structure (e.g., row and column context) to extract semantics among the cell entries.
- the Cell2Vec model may be based on Word2Vec, trained with all the cell entries in a row treated as context words. That is, the training is performed on the output of the cell-masking modules 120 . Word2Vec is described in Tomas Mikolov, Kai Chen, Gregory S.
- the Cell2Vec module 140 predicts target cells for embedded output (e.g., table 150 of FIG. 1 ) using context.
- target cells for embedded output e.g., table 150 of FIG. 1
- block 322 where the predicting considers row context using all the cell entries in the row of the target cell.
- block 323 Another example is illustrated by block 323 , where the predicting considers column context using a next N column entries (that is, entries in the column) relative to the column of the target cell.
- the Cell2Vec module 140 continues to predict until all target cells have been predicted and the corresponding embeddings of vectors have been made. At this point, a complete output table 150 having embeddings is created.
- TableNN module 160 tables contain data stored in cells, organized in rows and columns. Cell embeddings have been generated for every cell in the table. As illustrated in FIG. 1 , the cell and context embedding 155 are applied to the TableNN 160 module. The terms cell and cell-embedding are used interchangeably, since at this point, the cells contain cell embeddings. In block 330 of FIG. 3 , the TableNN module 160 , for a cell in the table, extracts a sliced table (i.e., less than all of the table 150 of FIG. 1 ) containing the cell and adjacent cells.
- a sliced table i.e., less than all of the table 150 of FIG. 1
- ⁇ t i,j b +( c i,j ,U t i,j ) W T (1)
- t (1, . . . , m+k)
- ⁇ t i,j is the linear transform output for each context element U t i,j and cell c i,j pair
- W ⁇ 1 ⁇ 2D is the learnable parameter (D is the cell embedding length) and b is the bias.
- Y i,j g fc ( b fc +( c i,j , ⁇ i,j ) W fc ), (4) where g fc , b fc and W fc are the activation function (e.g., ReLU, which is a non-linear activation function that is used in multi-layer neural networks or deep neural networks), bias, and learnable weights for the fully connected neural network respectively. Finally, Y i,j is processed through a log softmax function to calculate the class probabilities for the pre-defined classes 344 .
- ReLU ReLU
- a new framework is proposed to build a neural network-based model to predict the category of the cells using the context surrounding the cell entries.
- the proposed Cell-Masking tokenization method along with the Cell2Vec contextual embedding provides the highest performance. Also, attention-based modeling improves the header prediction accuracy significantly.
- the present invention may be a system, a method, and/or a computer program product.
- the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
- the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
- the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- SRAM static random access memory
- CD-ROM compact disc read-only memory
- DVD digital versatile disk
- memory stick a floppy disk
- a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
- a computer readable storage medium is not to be construed as being transitory signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire.
- Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
- the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
- Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer or entirely on the remote computer or server.
- the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
- electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute the computer readable program instructions by utilizing state information of the computer readable program instructions to personalize the electronic circuitry, in order to perform aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
- These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
- the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Molecular Biology (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Biomedical Technology (AREA)
- Probability & Statistics with Applications (AREA)
- Machine Translation (AREA)
Abstract
Description
ωt i,j =b+(c i,j ,U t i,j)W T (1)
where t=(1, . . . , m+k), ωt i,j is the linear transform output for each context element Ut i,j and cell ci,j pair, W∈ 1×2D is the learnable parameter (D is the cell embedding length) and b is the bias.
ζi,j=αi,j U i,j, (3)
where ζi,j ∈ D is the vector representing the context contribution. ζi,j is then concatenated with ci,j and the final output is processed through a fully connected neural network.
Y i,j =g fc(b fc+(c i,j,ζi,j)W fc), (4)
where gfc, bfc and Wfc are the activation function (e.g., ReLU, which is a non-linear activation function that is used in multi-layer neural networks or deep neural networks), bias, and learnable weights for the fully connected neural network respectively. Finally, Yi,j is processed through a log softmax function to calculate the class probabilities for the
Claims (18)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/483,989 US12254265B2 (en) | 2021-09-24 | 2021-09-24 | Generating unique word embeddings for jargon-specific tabular data for neural network training and usage |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/483,989 US12254265B2 (en) | 2021-09-24 | 2021-09-24 | Generating unique word embeddings for jargon-specific tabular data for neural network training and usage |
Publications (2)
Publication Number | Publication Date |
---|---|
US20230097150A1 US20230097150A1 (en) | 2023-03-30 |
US12254265B2 true US12254265B2 (en) | 2025-03-18 |
Family
ID=85718161
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US17/483,989 Active US12254265B2 (en) | 2021-09-24 | 2021-09-24 | Generating unique word embeddings for jargon-specific tabular data for neural network training and usage |
Country Status (1)
Country | Link |
---|---|
US (1) | US12254265B2 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP4478240A1 (en) * | 2023-06-12 | 2024-12-18 | Siemens Aktiengesellschaft | Method and system for automatic cell filling in a product specification table |
Citations (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6169990B1 (en) | 1996-03-02 | 2001-01-02 | University Of Strathclyde | Databases |
US6470306B1 (en) | 1996-04-23 | 2002-10-22 | Logovista Corporation | Automated translation of annotated text based on the determination of locations for inserting annotation tokens and linked ending, end-of-sentence or language tokens |
US20050027717A1 (en) | 2003-04-21 | 2005-02-03 | Nikolaos Koudas | Text joins for data cleansing and integration in a relational database management system |
US20060100974A1 (en) | 2004-10-22 | 2006-05-11 | International Business Machines Corporation | Visual structuring of multivariable data |
US20080059152A1 (en) | 2006-08-17 | 2008-03-06 | Neustar, Inc. | System and method for handling jargon in communication systems |
US20080082834A1 (en) * | 2006-09-29 | 2008-04-03 | Protegrity Corporation | Meta-complete data storage |
US20090164427A1 (en) | 2007-12-21 | 2009-06-25 | Georgetown University | Automated forensic document signatures |
US7664759B2 (en) | 2004-05-18 | 2010-02-16 | Hewlett-Packard Development Company, L.P. | Method and system for storing self-descriptive tabular data with alphanumeric and binary values |
US7777125B2 (en) | 2004-11-19 | 2010-08-17 | Microsoft Corporation | Constructing a table of music similarity vectors from a music similarity graph |
US20100312769A1 (en) | 2009-06-09 | 2010-12-09 | Bailey Edward J | Methods, apparatus and software for analyzing the content of micro-blog messages |
US20110125578A1 (en) | 2000-04-04 | 2011-05-26 | Aol Inc. | Filtering system for providing personalized information in the absence of negative data |
US7966174B1 (en) * | 1998-12-07 | 2011-06-21 | At&T Intellectual Property Ii, L.P. | Automatic clustering of tokens from a corpus for grammar acquisition |
US20150052084A1 (en) | 2013-08-16 | 2015-02-19 | Kabushiki Kaisha Toshiba | Computer generated emulation of a subject |
US20170286398A1 (en) | 2016-03-29 | 2017-10-05 | Geoffrey Hodgson Hunt | Method to resolve the meaning of a body of natural language text using artificial intelligence analysis in combination with semantic and contextual analysis |
US20170300471A1 (en) | 2014-09-30 | 2017-10-19 | Hewlett-Packard Development Company, L.P. | Specialized language identification |
US9911003B2 (en) | 2015-11-01 | 2018-03-06 | International Business Machines Corporation | Dynamic data masking of post-output database data |
US10229154B2 (en) | 2013-07-18 | 2019-03-12 | International Business Machines Corporation | Subject-matter analysis of tabular data |
US10394803B2 (en) | 2015-11-13 | 2019-08-27 | International Business Machines Corporation | Method and system for semantic-based queries using word vector representation |
US20200089769A1 (en) | 2018-01-04 | 2020-03-19 | Facebook, Inc. | Consumer Insights Analysis Using Word Embeddings |
US20200118555A1 (en) | 2018-10-10 | 2020-04-16 | N3, Llc | Semantic jargon |
US10657332B2 (en) | 2017-12-21 | 2020-05-19 | Facebook, Inc. | Language-agnostic understanding |
US20200175390A1 (en) * | 2018-11-29 | 2020-06-04 | International Business Machines Corporation | Word embedding model parameter advisor |
US20200210520A1 (en) * | 2018-12-26 | 2020-07-02 | Nokia Solutions And Networks Oy | Determination of field types in tabular data |
US20210182476A1 (en) * | 2018-06-01 | 2021-06-17 | Palantir Technologies Inc. | Transformation in tabular data cleaning tool |
US20220309087A1 (en) * | 2021-03-29 | 2022-09-29 | Google Llc | Systems and methods for training language models to reason over tables |
-
2021
- 2021-09-24 US US17/483,989 patent/US12254265B2/en active Active
Patent Citations (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6169990B1 (en) | 1996-03-02 | 2001-01-02 | University Of Strathclyde | Databases |
US6470306B1 (en) | 1996-04-23 | 2002-10-22 | Logovista Corporation | Automated translation of annotated text based on the determination of locations for inserting annotation tokens and linked ending, end-of-sentence or language tokens |
US7966174B1 (en) * | 1998-12-07 | 2011-06-21 | At&T Intellectual Property Ii, L.P. | Automatic clustering of tokens from a corpus for grammar acquisition |
US20110125578A1 (en) | 2000-04-04 | 2011-05-26 | Aol Inc. | Filtering system for providing personalized information in the absence of negative data |
US20050027717A1 (en) | 2003-04-21 | 2005-02-03 | Nikolaos Koudas | Text joins for data cleansing and integration in a relational database management system |
US7664759B2 (en) | 2004-05-18 | 2010-02-16 | Hewlett-Packard Development Company, L.P. | Method and system for storing self-descriptive tabular data with alphanumeric and binary values |
US20060100974A1 (en) | 2004-10-22 | 2006-05-11 | International Business Machines Corporation | Visual structuring of multivariable data |
US7777125B2 (en) | 2004-11-19 | 2010-08-17 | Microsoft Corporation | Constructing a table of music similarity vectors from a music similarity graph |
US20080059152A1 (en) | 2006-08-17 | 2008-03-06 | Neustar, Inc. | System and method for handling jargon in communication systems |
US20080082834A1 (en) * | 2006-09-29 | 2008-04-03 | Protegrity Corporation | Meta-complete data storage |
US20090164427A1 (en) | 2007-12-21 | 2009-06-25 | Georgetown University | Automated forensic document signatures |
US20100312769A1 (en) | 2009-06-09 | 2010-12-09 | Bailey Edward J | Methods, apparatus and software for analyzing the content of micro-blog messages |
US10229154B2 (en) | 2013-07-18 | 2019-03-12 | International Business Machines Corporation | Subject-matter analysis of tabular data |
US20150052084A1 (en) | 2013-08-16 | 2015-02-19 | Kabushiki Kaisha Toshiba | Computer generated emulation of a subject |
US20170300471A1 (en) | 2014-09-30 | 2017-10-19 | Hewlett-Packard Development Company, L.P. | Specialized language identification |
US9911003B2 (en) | 2015-11-01 | 2018-03-06 | International Business Machines Corporation | Dynamic data masking of post-output database data |
US10394803B2 (en) | 2015-11-13 | 2019-08-27 | International Business Machines Corporation | Method and system for semantic-based queries using word vector representation |
US20170286398A1 (en) | 2016-03-29 | 2017-10-05 | Geoffrey Hodgson Hunt | Method to resolve the meaning of a body of natural language text using artificial intelligence analysis in combination with semantic and contextual analysis |
US10657332B2 (en) | 2017-12-21 | 2020-05-19 | Facebook, Inc. | Language-agnostic understanding |
US20200089769A1 (en) | 2018-01-04 | 2020-03-19 | Facebook, Inc. | Consumer Insights Analysis Using Word Embeddings |
US20210182476A1 (en) * | 2018-06-01 | 2021-06-17 | Palantir Technologies Inc. | Transformation in tabular data cleaning tool |
US20200118555A1 (en) | 2018-10-10 | 2020-04-16 | N3, Llc | Semantic jargon |
US20200175390A1 (en) * | 2018-11-29 | 2020-06-04 | International Business Machines Corporation | Word embedding model parameter advisor |
US20200210520A1 (en) * | 2018-12-26 | 2020-07-02 | Nokia Solutions And Networks Oy | Determination of field types in tabular data |
US10817657B2 (en) | 2018-12-26 | 2020-10-27 | Nokia Solutions And Networks Oy | Determination of field types in tabular data |
US20220309087A1 (en) * | 2021-03-29 | 2022-09-29 | Google Llc | Systems and methods for training language models to reason over tables |
Non-Patent Citations (35)
Title |
---|
0. LEVY et al.; "Linguistic regularities in sparse and explicit word representations"; In Proceedings of the Eighteenth Conference on Computational Natural Language Learning, CoNLL, Baltimore, Maryland, USA; Jun. 26-27, 2014; pp. 171-180. |
0. LEVY et al.; "Neural word embedding as implicit matrix factorization"; In Advances In Neural Information Processing Systems 27: Annual Conference on Neural Information Processing Systems, Montreal, Quebec, Canada; Dec. 8-13, 2014; pp. 2177-2185. |
A. Mnih et al.; "Learning word embeddings effciently with noise-contrastive estimation"; In Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013; Proceedings of a meeting held Dec. 5-8, 2013, Lake Tahoe, Nevada, United States; pp. 2265-2273. |
Arik, Sercan O. et al., "TabNet: Attentive Interpretable Tabular Learning", arXiv preprint arXiv:1908.07442, 2019 , 9 pgs. |
Bahdanau, Dzmitry, et al., "Neural Machine Translation By Jointly Learning To Align And Translate", ICLR 2015, arXiv:1409.0473v7 [cs.CL], May 19, 2016, 15 pgs. |
Bhagavatula et al., "TabEL: entity linking in web tables", International Semantic Web Conference, 2015, 425—441, Springer, <http://iswc2015.semanticweb.org/sites/iswc2015.semanticweb.org/files/93660385.pdf>, 16 pages. |
Chen, Jiaoyan, et al., "Learning Semantic Annotations for Tabular Data", arXiv:1906.00781v1 [cs.DB], May 30, 2019, 7 pgs. |
Deng, Li, et al., "Table2Vec: Neural Word and Entity Embeddings for Table Population and Retrieval", Short Research Papers 2A:AI, Mining, and others, SIGIR '19, Jul. 21-25, 2019, 4 pgs. |
Devlin, Jacob, et al., "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding", Proceedings of NAACL-HLT 2019, pp. 4171-4186. |
Dreher et al ("Assisted query formulation using normalised word vector and dynamic ontological filtering", vol. 4027 LNAI, 2006, pp. 282-294, 7th International Conference on Flexible Query Answering Systems, FQAS 2006; Milan; Italy; Jun. 7, 2006 through Jul. 10, 2006; Code 67889). |
Efthymiou, Vasilis, et al., "Matching Web Tables with Knowledge Base Entities: From Entity Lookups to Entity Embeddings", International Semantic Web Conference, 2017, 16 pages. |
Farrell, Tracie, et al., "On the use of Jargon and Word Embeddings to Explore Subculture within the Reddit's Manosphere", WebSci '20, Jul. 6-10, 2020, 10 pgs. |
Ghasemi-Gol, Majid, et al., "TabVec: Table Vectors for Classification of Web Tables", Conference '17, Jul. 2017; arXiv:1802.06290v1 [cs.IR]) Feb. 17, 2018, 9 pgs. |
Herzig, Jonathan, et al., "TAPAS: Weakly Supervised Table Parsing via Pre-training", ar:Xiv:2004.02349v2 [cs.IR], Apr. 21, 2020, 14 pgs. |
ISO 6346 Container Type Codes, https://datahub.io/core/iso-container-codes#data, Aug. 2021, 14 pgs. |
J. Pennington et al.; "Glove: Global vectors for word representation"; In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014, Oct. 25-29, 2014, Doha, Qatar; A meeting of SIGDAT, a Special Interest Group of the ACL; pp. 1532-1543. |
Limaye, Girija, et al., "Annotating and Searching Web Tables Using Entities, Types and Relationships", PVLDB, 3:1338-1347, Sep. 2010. |
Mikolov, Tomas, et al., "Efficient Estimation of Word Representation in Vector Space", ar:Xiv.1301.3781v3 [cs.CL], Sep. 7, 2013, 12 pgs. |
Mulwad et al., "Semantic Message Passing for Generating Linked Data from Tables", 12th International Semantic Web Conference, Sydney, Oct. 2013, 19 pages. |
Mulwad et al., "Using linked data to interpret tables", Proceedings of the the First International Workshop on Consuming Linked Data, 2010, <https://ebiquity.umbc.edu/_file_directory_/papers/524.pdf>, 12 pages. |
Pham et al., "Semantic labeling: a domain-independent approach", International Semantic Web Conference, 2016, 446—462, Springer, <https://usc-isi-i2.github.io/papers/pham16-iswc.pdf>, 16 pages. |
Ren et al., "Likelihood Ratios for Out-of-Distribution Detection", 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, Canada, arXiv:1906.02845v2 [stat. ML] Dec. 5, 2019, 21 pages. |
S. Arora et al., "Random Walks on Context Spaces: Towards an Explanation of the Mysteries of Semantic Word Embeddings", arXiv:1502.03520v3 [cs.LG] Apr. 21, 2015, 28 pages. |
S. Arora et al.; "Rand-Walk: A latent variable model approach to word embeddings"; arXiv:1502.03520v5 [cs.LG] Oct. 14, 2015, 23 pages. |
Sennich, Rico, et al., "Neural Machine Translation of Rare Words with Subword Units", ar:Xiv:1508.07909v5 [cs.CL}, Jun. 10, 2016, 11 pgs. |
Sun, Baohua, et al., "SuperTML: Two-Dimensional Word Embedding for the Precognition on Structured Tabular Data", Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2019, 9 pgs. |
Sutskever, Ilya, et al., "Sequence to sequence Learning with Neural Networks", ar:Xiv:1409.3215v3 [cs CL], Dec. 14, 2014, 9 pgs. |
T. Mikolov et al.; "Distributed representations of words and phrases and their compositionality"; In Advances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013; Proceedings of a meeting held Dec. 5-8, 2013, Lake Tahoe, Nevada, United States; pp. 3111-3119. |
T. Mikolov et al.; "Exploiting similarities among languages for machine translation"; CoRR, abs/1309.4168; 2013; whole document (10 pages). |
T. Mikolov; "word2vec: Tool for computing continuous distributed representations of words"; https://code.google.com/p/word2vec; 2015; whole document (4 pages). |
Takeoka et al., "An efficient probabilistic approach for semantically annotating table", The Thirty-Third AAAI Conference on Artificial Intelligence (AAAI-19), 2019, 8 pages. |
Venetis et al., "Recovering semantics of tables on the web", Proceedings of the VLDB9, vol. 4, Issue 9, 11 pages, Published Jun. 1, 2011, <https://doi.org/10.14778/2002938.2002939>. |
Y. Bengio et al.; "A neural probabilistic language model"; Journal of Machine Learning Research, 3; 2003; pp. 1137-1155. |
Ying, Pengcheng et al., "TaBert: Pretraining for Joint Understanding of Textual and Tabular Data", ACL 2010, arXiv:2005.08314v1 [cs.CL], May 17, 2020, 15 pgs. |
Yu, Mo, et al., "Factor-based Compositional Embedding Models", NIPS Workshop on Learning Semantics, 2014, 5 pgs. |
Also Published As
Publication number | Publication date |
---|---|
US20230097150A1 (en) | 2023-03-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7282940B2 (en) | System and method for contextual retrieval of electronic records | |
Dashtipour et al. | Exploiting deep learning for Persian sentiment analysis | |
US10095690B2 (en) | Automated ontology building | |
Nguyen et al. | Recurrent neural network-based models for recognizing requisite and effectuation parts in legal texts | |
Pham et al. | End-to-end recurrent neural network models for vietnamese named entity recognition: Word-level vs. character-level | |
Chauhan et al. | Aspect based sentiment analysis using deep learning approaches: A survey | |
US10585989B1 (en) | Machine-learning based detection and classification of personally identifiable information | |
US12141697B2 (en) | Annotating customer data | |
US11797842B2 (en) | Identifying friction points in customer data | |
Nasr et al. | Building sentiment analysis model using Graphlab | |
Köksal et al. | Improving automated Turkish text classification with learning‐based algorithms | |
Sitender et al. | Sanskrit to universal networking language EnConverter system based on deep learning and context-free grammar | |
US12254265B2 (en) | Generating unique word embeddings for jargon-specific tabular data for neural network training and usage | |
Mishra et al. | Explainability for NLP | |
Li et al. | Tagdeeprec: tag recommendation for software information sites using attention-based bi-lstm | |
Sabty et al. | Techniques for named entity recognition on arabic-english code-mixed data | |
Priyadarshi et al. | A study on the performance of recurrent neural network based models in Maithili part of speech tagging | |
Corredera Arbide et al. | Affective computing for smart operations: a survey and comparative analysis of the available tools, libraries and web services | |
Dahiya et al. | Curriculum learning strategies for Hindi-English code-mixed sentiment analysis | |
Zaikis et al. | Dacl: A domain-adapted contrastive learning approach to low resource language representations for document clustering tasks | |
Tun et al. | Stock article title sentiment-based classification using PhoBERT | |
CN113822018A (en) | Entity Relation Joint Extraction Method | |
Vo | SE4SA: a deep syntactical contextualized text representation learning approach for sentiment analysis | |
He | An intelligent algorithm for fast machine translation of long English sentences | |
Singh et al. | Intelligent Text Mining Model for English Language Using Deep Neural Network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
AS | Assignment |
Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW YORK Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:AGRAWAL, BHAVNA;KHABIRI, ELHAM;LI, YINGJIE;AND OTHERS;SIGNING DATES FROM 20210928 TO 20211021;REEL/FRAME:057874/0328 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |