Skip to content
This repository was archived by the owner on Jan 19, 2019. It is now read-only.
198 changes: 198 additions & 0 deletions deep_qa/data/instances/text_classification/frame_instance.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,198 @@
from typing import Dict, List

import numpy
from overrides import overrides

from ..instance import TextInstance, IndexedInstance
from ...data_indexer import DataIndexer

# TODO PR request for having these in the json as an application specific content
# the slotnames can vary according to different end applications, e.g., a HowTo tuple, OpenIE tuple ...
SLOTNAMES_ORDERED = ["agent", "beneficiary", "causer", "context", "definition", "event",
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If you want to have these global variables, you could consider moving them inside FrameInstance but above init. This would then make them class attributes, which would be accessible in your classmethods.

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

If you want to always have a fixed set of slots that every dataset uses, where your input includes a one-hot vector for which slot you're trying to fill, or something like that, this should work. The model's weights will be tied to particular slots, so a model trained on one slot configuration will not possibly generalize to any other slot configuration.

If you want to make this more modular, you could consider using a separate namespace in the data indexer to keep track of this for you, instead of hard-coding it here. Something similar is done with tag labels here.

Copy link
Copy Markdown
Author

@nikett nikett Jun 7, 2017

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The reason to have this configurable was to be able to train any variant of the input with the same code, by passing a list with the config file upfront. This configuration allows ordering the input on the fly for any types of slots.

It appears that there is no possibility to specify these list of slotnames in the config file, so, perhaps having this as a constant may be better than passing through a function. The data indexer does not really do any book keeping because the frame_instance takes care of it.

Am I thinking incorrectly?

Copy link
Copy Markdown
Contributor

@matt-gardner matt-gardner Jun 7, 2017

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oh, yeah, your preprocess scripts putting everything into this format will take care of that, I think.

Another consideration is that you'll be essentially wasting computation and modeling capacity on all of the un-filled slots by filling them with "unk". You could either: (1) leave them empty, and let them be masked in whatever model computation you do later. This will keep the model from wasting its capacity trying to figure out that "unk" means "unfilled". Or (2) instead of having a fixed-sized input here, you could have a key-value input, where you give a list of (slot, value) pairs as input to the model. In this setting, it's also a lot easier to think about how to pass in the query - it's just another (slot, value) pair, just with a "QUERY" value, or something similar. Then instead of learning what is essentially an embedding of each slot type in the weights of a softmax layer, you can explicitly learn an embedding for them, and have that be part of your model.

Happy to talk about this more in person if it's hard to follow what I'm suggesting here.

"finalloc", "headverb", "initloc", "input", "output", "manner",
"patient", "resultant", "timebegin", "timeend", "temporal", "hierarchical",
"similar", "contemporary", "enables", "mechanism", "condition", "purpose",
"cause", "openrel", "participant"]
UNKNOWN_SLOTVAL = "unk" # making an open world assumption, we do not observe all the values
QUES_SLOTVAL = "ques" # this slot in the frame must be queried/completed.
MAX_PHRASE_LEN = 6


class FrameInstance(TextInstance):

"""
A FrameInstance is a kind of TextInstance that has text in multiple slots. This generalizes a FrameInstance.
"""
def __init__(self,
padded_slots: List[str],
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

There is no padding done here, so I wouldn't put padded in this name.

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oh, looks like I'm wrong. This would be more clear if there were some comment in the docstring about what this represents. Also, this will run into some naming collision with padding as done in the indexed instances.

Copy link
Copy Markdown
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed to sparse and dense frame, instead of unpadded and padded slots.

phrase_in_queried_slot: str=None): # output label is a phrase
super(FrameInstance, self).__init__(phrase_in_queried_slot)
self.text = padded_slots # "event:plant absorb water###participant:water###agent:plant" TAB "agent:plant"

def __str__(self):
return 'FrameInstance( [' + ',\n'.join(self.text) + '] , ' + str(self.label) + ')'

@overrides
def words(self) -> Dict[str, List[str]]:
# Accumulate words from each slot's phrase.
# Label is also a phrase, so additionally accumulate words from label
words = []
for phrase in self.text: # phrases
phrase_words = self._words_from_text(phrase)
words.extend(phrase_words['words'])
label_words = self._words_from_text(self.label)
words.extend(label_words['words'])
return {"words": words}

@staticmethod
def query_slot_from(slot_as_string: str,
sparse_given_frame: Dict[str, str],
kv_separator: str=":"):
"""
:param slot_as_string: "participant:water"
:param sparse_given_frame: If the expected slot name is given in the query
but its value is not, then pick the value from the sparse_given_frame
:param kv_separator: typically colon separated
:return: name=participant, val=water
"""
slot_name_val = slot_as_string.split(kv_separator)
# Suppose slot_as_string is: participant (i.e. no value is specified)
# this is assumed as participant:BLANK_VALUE, if we cannot look it up in the partial frame.
if len(slot_name_val) == 1:
slot_name_val = (slot_as_string + ":" +
sparse_given_frame.get(slot_name_val[0], '')).split(kv_separator)
return {'name': slot_name_val[0], 'val': slot_name_val[1]}

@staticmethod
def unpack_input(frame_as_string: str,
kv_separator: str="\t"):
"""
:param frame_as_string: "event:plant absorb water###participant:water" TAB "participant:water"
:param kv_separator: typically TAB separated partial frame and query
:return: event:plant absorb water###participant:water, and query: participant:water
Both event and query will be lowercased
"""
# No information loss in lower-casing, and simplifies matching.
partialframe_query = frame_as_string.lower().split(kv_separator)
if len(partialframe_query) != 2:
raise RuntimeError("Unexpected number (not 2) of fields in frame: " + frame_as_string)
return {'content': partialframe_query[0], 'query': partialframe_query[1]}

@staticmethod
def given_slots_from(slots_csv: str,
values_separator: str="###",
kv_separator: str=":"):
"""
:param slots_csv: event:plant absorb water###participant:water
:param values_separator: typically "###"
:param kv_separator: typically ":"
:return: map of slotnames -> slot phrase [event -> plant absorb water , participant -> water]
"""
return dict(map(lambda x: x.split(kv_separator), slots_csv.split(values_separator)))
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Could you add a comment describing this function? It's not clear immediately what it does. You could consider re-writing this like:

key_value_list = slots_csv.split(values_separator)
key_value_tuples = [pair.split(key_value_separator) for x in key_value_list]
key_value_dict = {key: value for key, value in key_value_tuples}


@staticmethod
def padded_slots_from(sparse_frame: Dict[str, str],
query_slotname: str):
"""
Performs two types of padding:
i) unobserved slots are filled with self.unknown_slotval
ii) query slot is masked with self.unknown_queryval
The order of slots strictly follows from SLOTNAMES_ORDERED.
:param sparse_frame:
slotnames -> slot phrase [event -> plant absorb water , participant -> water]
:param query_slotname:
participant
:return: [plant absorb water, ques, unk, unk, ...]
"""
slots = []
for slotname in SLOTNAMES_ORDERED:
if slotname == query_slotname: # query hence masked
slots.append(QUES_SLOTVAL)
elif slotname in sparse_frame: # observed hence as-is
slots.append(sparse_frame[slotname])
else: # unobserved hence padded
slots.append(UNKNOWN_SLOTVAL)
return slots

@classmethod
@overrides
def read_from_line(cls, line: str):
"""
Reads a FrameInstance from a line. The format is:
frame represented as list of <role:role value phrase of maxlen 5> TAB <label>
e.g., from
event:plant absorb water###participant:water###agent:plant###finalloc:soil
to
["plant", "unk", "unk", "unk", "unk", "plant absorb water",
"soil", "unk", "unk", "unk", "unk", "unk",
"unk", "unk", "unk", "unk", "unk", "unk",
"unk", "unk", "unk", "unk", "unk", "unk",
"unk", "unk", "water"]
Provides ordering (input can be composed of slots in arbitrary order)
and sparseness flexibility (only a few slots can be mentioned in the input).
"""
# Extract the query slot name and expected value
# e.g. from, participant:water, extract the expected slot value "water"
unpacked_input = cls.unpack_input(line)
given_slots = cls.given_slots_from(unpacked_input['content'])
query_slot = cls.query_slot_from(unpacked_input['query'], given_slots)
padded_slots = cls.padded_slots_from(given_slots, query_slot['name'])
return cls(padded_slots, phrase_in_queried_slot=query_slot['val'])

@overrides
def to_indexed_instance(self, data_indexer: DataIndexer):
# A phrase in a slot, is converted from list of words to list of wordids.
# This is repeated for every slot, hence a list of list of wordids/integers.
indices_slotvals = [self._index_text(phrase, data_indexer) for phrase in self.text]
# The label is a phrase, and is converted from list of words to list of wordids.
indices_label = self._index_text(self.label, data_indexer)
return IndexedFrameInstance(indices_slotvals, indices_label)


class IndexedFrameInstance(IndexedInstance):
"""
Ensures that a phrase in every slot, and the label (also a phrase) are padded to be of a fixed maxlen.
Max length of a phrase is 6 (configurable), pad phrases with fewer words; if it exceeds 6 then truncate.
"""
def __init__(self, word_indices: List[List[int]], label):
"""
:param word_indices: One list of ints make up a slotvalue because a slotvalue is a phrase,
and so every word of the phrase is identified with an int id.
:param label: phrase, hence a list of ints.
"""
super(IndexedFrameInstance, self).__init__(label)
self.word_indices = word_indices

@classmethod
@overrides
def empty_instance(cls):
return IndexedFrameInstance([], label=None)

@overrides
def get_padding_lengths(self) -> Dict[str, int]:
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Take a look at how get_padding_lengths is implemented here - this similar to your case, where you have a set of slots which need padding. In this method, we only care about the max length of a slot.
https://github.com/allenai/deep_qa/blob/master/deep_qa/data/instances/text_classification/tuple_instance.py#L94

Then, you can see here that the pad method expects to get back this dictionary - this answers the question in your comment above.
https://github.com/allenai/deep_qa/blob/master/deep_qa/data/instances/text_classification/tuple_instance.py#L100

Implementing these two methods like this will allow you to remove the MAX_SLOT_LEN global variable and have the padding for the slots set automatically from your training data.

# The parent class somehow understands the semantics of these dictionary keys.
# would be better if these keys were more explicit, e.g., as enumerations.
return {'num_sentence_words': MAX_PHRASE_LEN}
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yeah, there's no need to hard-code a length here, and that parameter is better specified in the model you build, anyway. This will actually break the code we have that tries to be smart about letting you swap out word representations. It looks like the tuple instance code that Mark linked to also will break the word representation stuff that we have. Instead, what you should do is something like this.

@DeNeutoy, I think we need way better documentation around how the whole data pipeline works. I think it does some really nice things, but people have a hard time understanding it and how to use it correctly.

Copy link
Copy Markdown
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I now understand. So something like this?

for key in slot_word_lengths:
            lengths[key] = len(key)
return lengths```

Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yeah, similar. I would do something like this:

all_slot_lengths = [self._get_word_sequence_lengths(slot_indices) for slot_indices in self.word_indices]
lengths = {}
for key in slot_lengths[0]:
    lengths[key] = max(slot_lengths[key] for slot_lengths in all_slot_lengths)
return lengths


@overrides
def pad(self, padding_lengths: Dict[str, int]):
"""
Pads (or truncates) all slot values to the maxlen
Input: (phrases corresponding to each slot, the number of slots is fixed.)
e.g., ["1000", "1", "1", "1", "1", "1 2 3",..]
Note: these are arrays over phrase word ids.
Output: (padded phrases, as phrases are composed of variable number of words)
e.g., ["1000 0 0 0 0", "1 0 0 0 0", "1 0 0 0 0", "1 0 0 0 0", "1 0 0 0 0", "1 2 3 0 0",..]
Note: padding is fixed length, anything larger or small will be pruned. Phrases are truncated from left.
"""
truncate_from_right = False
self.word_indices = [self.pad_word_sequence(indices, padding_lengths, truncate_from_right)
for indices in self.word_indices]
self.label = self.pad_word_sequence(self.label, padding_lengths, truncate_from_right)

@overrides
def as_training_data(self):
# The frame and the label must be numpy matrix and array respectively
frame_as_matrix = numpy.asarray(self.word_indices, dtype='int32')
label_as_list = numpy.asarray(self.label, dtype='int32')
return frame_as_matrix, label_as_list
86 changes: 86 additions & 0 deletions tests/data/instances/text_classification/frame_instance_test.py
Original file line number Diff line number Diff line change
@@ -0,0 +1,86 @@
# pylint: disable=no-self-use,invalid-name
from deep_qa.data import DataIndexer

from deep_qa.data.instances.text_classification.frame_instance import FrameInstance, IndexedFrameInstance
from tests.common.test_case import DeepQaTestCase


class TestFrameInstance(DeepQaTestCase):

def setUp(self):
super(TestFrameInstance, self).setUp()
# Example of a typical input
self.line = "event:plant absorb water###" \
"participant:water###" \
"agent:plant###" \
"finalloc:soil"\
+ "\t" + "finalloc:soil"
self.line_with_no_label_val = "event:plant absorb water###" \
"participant:water###" \
"agent:plant###" \
"finalloc:soil" \
+ "\t" + "finalloc"
self.padded_slots = ['plant', 'unk', 'unk', 'unk', 'unk',
'plant absorb water', 'ques', 'unk', 'unk',
'unk', 'unk', 'unk', 'unk', 'unk', 'unk',
'unk', 'unk', 'unk', 'unk', 'unk', 'unk', 'unk',
'unk', 'unk', 'unk', 'unk', 'water']
self.data_indexer = DataIndexer()
for word in ['plant', 'unk', 'absorb', 'ques', 'water', 'soil']:
self.data_indexer.add_word_to_index(word)

def test_convert_instance_to_indexed_instance(self):
instance = FrameInstance.read_from_line(self.line)
indexed_instance = instance.to_indexed_instance(self.data_indexer)
assert indexed_instance.label == [self.data_indexer.get_word_index('soil')]

def test_convert_instance_no_label_value_to_indexed_instance(self):
instance = FrameInstance.read_from_line(self.line_with_no_label_val)
indexed_instance = instance.to_indexed_instance(self.data_indexer)
assert indexed_instance.label == [self.data_indexer.get_word_index('soil')]

def test_slots_unwrap_correctly(self):
instance = FrameInstance.read_from_line(self.line)
# what we construct
machine_label = instance.label
machine_slot_values = instance.text
# what we expect
expected_label = "soil"
# do they match?
assert machine_label == expected_label
assert machine_slot_values == self.padded_slots
Copy link
Copy Markdown
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

The label here is what you're trying to predict. You want to have the model output "soil". But it looks like there's nothing in this instance that tells the model what slot you're trying to query. How would a model know which slot to fill, when it's trying to predict "soil"?

Copy link
Copy Markdown
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Very good point. I was hoping to include a one-hot vector depicting the slot to fill. I thought that by specifying a special tag "ques", I could inform the model eventually. You are probably right, that in the data instance, my input could explicitly include a one-hot vector. I don't know how to do that so I would ask Mark.


def test_slots_no_label_value_unwrap_correctly(self):
instance = FrameInstance.read_from_line(self.line_with_no_label_val)
# what we construct
machine_label = instance.label
machine_slot_values = instance.text
# what we expect
expected_label = "soil"
# do they match?
assert machine_label == expected_label
assert machine_slot_values == self.padded_slots

def test_words_from_frame_aggregated_correctly(self):
instance = FrameInstance.read_from_line(self.line)
assert len(instance.words()['words']) == 30

def test_words_from_no_label_value_frame_aggregated_correctly(self):
instance = FrameInstance.read_from_line(self.line_with_no_label_val)
assert len(instance.words()['words']) == 30


class TestIndexedFrameInstance(DeepQaTestCase):

def test_words_from_frame_aggregated_correctly(self):
indexed_instance = IndexedFrameInstance([[1000], [1, 2, 3, 4, 5, 6, 7, 8],
[1, 2, 3]], [1, 2, 3])
# unpadded label should be read correctly.
assert indexed_instance.label == [1, 2, 3]
padding_lengths = indexed_instance.get_padding_lengths()
assert padding_lengths['num_sentence_words'] == 6
indexed_instance.pad(padding_lengths)
assert indexed_instance.label == [1, 2, 3, 0, 0, 0]
assert indexed_instance.word_indices == [[1000, 0, 0, 0, 0, 0],
[1, 2, 3, 4, 5, 6],
[1, 2, 3, 0, 0, 0]]