Merge branch 'master' into dev/wutlin

This commit is contained in:
Wutao Lin 2019-08-13 16:03:48 +08:00
Родитель 516376ebe9 58ad563a23
Коммит 1b4995c639
63 изменённых файлов: 6397 добавлений и 509 удалений

1
.gitignore поставляемый
Просмотреть файл

@ -5,4 +5,5 @@
*.vs*
dataset/GloVe/
dataset/20_newsgroups/
dataset/SST-2/
models/

Просмотреть файл

@ -18,6 +18,64 @@ contact [opencode@microsoft.com](mailto:opencode@microsoft.com) with any additio
- Contribute Model to **Model Zoo**
- We encourage everyone to contribute their NLP models (namely JSON configuration files). Please follow the structure in model_zoo to create a pull request.
- Contribute Block to **Block Zoo**
- We encourage everyone to improve this toolkit by contributing code, such as customized Blocks. So other users can further benefit from these new Blocks.
We encourage everyone to improve this toolkit by contributing code, such as customized Blocks. So other users can further benefit from these new Blocks.
For adding a new block to NeuronBlocks, you need *Three steps*(take [BiLSTM block](https://github.com/microsoft/NeuronBlocks/blob/master/block_zoo/BiLSTM.py) for example):
- Define the new block's Configuration class(BiLSTMConf class in BiLSTM block). The Configuration class should inheritance [Base Configuration Class](https://github.com/microsoft/NeuronBlocks/blob/master/block_zoo/BaseLayer.py) that define some necessary functions, and rewrite these functions.
We will give the details of these functions:
```bash
def default():
'''
Define the default hyper parameters for block, it will read the corresponding block hyper parameters in configuration json files firstly.
'''
def declare():
'''
Define things like "input_ranks" and "num_of_inputs", which are certain with regard to the block.
num_of_input is N(N>0) means this layer accepts N inputs;
num_of_input is -1 means this layer accepts any number of inputs;
The rank here is not the same as matrix rank:
For a scalar, its rank is 0;
For a vector, its rank is 1;
For a matrix, its rank is 2;
For a cube of numbers, its rank is 3.
if num_of_input > 0:
len(input_ranks) should be equal to num_of_input
elif num_of_input == -1:
input_ranks should be a list with only one element and the rank of all the inputs should be equal to that element.
'''
def inference():
'''
Inference things like output_dim, which may relies on defined hyper parameter or the block special operation.
'''
def verify():
'''
Define some necessary varification for your layer when we define the model.
'''
```
- Implement the new block's class(BiLSTM class in BiLSTM block). The block class should inheritance [Base Block Class](https://github.com/microsoft/NeuronBlocks/blob/master/block_zoo/BaseLayer.py) and rewrite __init__ and forward function.
```bash
def __init__():
'''
Define necessary attributions that would be used in block operation logic.
'''
def forward():
'''
Tensor operation logic.
'''
```
- Register the new block in block_zoo.
NeuronBlocks provides a script that can register new block automatically, and blocks contributors just focus on block logic.
*Tips: PROJECTROOT denotes the root directory of this project.*
```bash
cd PROJECT_ROOT
python register_block.py --block_name=new_block_name
```
*Tips: Before you contribute your code, we strongly suggest to verify that your improvements are valid by **[AUTOTEST](./autotest)**. We also encourage everyone to improve this autotest tool by contributing code, such as adding test tasks.*

Просмотреть файл

@ -13,7 +13,7 @@ import codecs
import pickle as pkl
from utils.common_utils import dump_to_pkl, load_from_pkl, get_param_num, get_trainable_param_num, \
transfer_to_gpu, transform_params2tensors
transfer_to_gpu, transform_params2tensors, get_layer_class, load_from_json, dump_to_json
from utils.philly_utils import HDFSDirectTransferer, open_and_move, convert_to_tmppath, \
convert_to_hdfspath, move_from_local_to_hdfs
from Model import Model
@ -22,8 +22,10 @@ from metrics.Evaluator import Evaluator
from utils.corpus_utils import get_batches
from core.StreamingRecorder import StreamingRecorder
from core.LRScheduler import LRScheduler
from settings import ProblemTypes
from settings import ProblemTypes, Setting as st
from block_zoo import Linear
from block_zoo import CRF
from losses.CRFLoss import CRFLoss
class LearningMachine(object):
@ -40,6 +42,7 @@ class LearningMachine(object):
device = 'GPU' if 'cuda' in emb_weight_device else 'CPU'
logging.info(
"The embedding matrix is on %s now, you can modify the weight_on_gpu parameter to change embeddings weight device." % device)
logging.info("="*100 + '\n' + "*"*15 + "Model Achitecture" + "*"*15)
logging.info(self.model)
#logging.info("Total parameters: %d; trainable parameters: %d" % (get_param_num(self.model), get_trainable_param_num(self.model)))
logging.info("Total trainable parameters: %d" % (get_trainable_param_num(self.model)))
@ -89,33 +92,18 @@ class LearningMachine(object):
def train(self, optimizer, loss_fn):
self.model.train()
if not self.conf.train_data_path.endswith('.pkl'):
train_data, train_length, train_target = self.problem.encode(self.conf.train_data_path, self.conf.file_columns,
self.conf.input_types, self.conf.file_with_col_header, self.conf.object_inputs, self.conf.answer_column_name, max_lengths=self.conf.max_lengths,
min_sentence_len = self.conf.min_sentence_len, extra_feature=self.conf.extra_feature,fixed_lengths=self.conf.fixed_lengths, file_format='tsv',
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers)
else:
train_pkl_data = load_from_pkl(self.conf.train_data_path)
train_data, train_length, train_target = train_pkl_data['data'], train_pkl_data['length'], train_pkl_data['target']
logging.info("="*100 + '\n' + "*"*15 + 'Prepare data for training' + "*"*15)
if not self.conf.valid_data_path.endswith('.pkl'):
valid_data, valid_length, valid_target = self.problem.encode(self.conf.valid_data_path, self.conf.file_columns,
self.conf.input_types, self.conf.file_with_col_header, self.conf.object_inputs, self.conf.answer_column_name, max_lengths=self.conf.max_lengths,
min_sentence_len = self.conf.min_sentence_len, extra_feature = self.conf.extra_feature,fixed_lengths=self.conf.fixed_lengths, file_format='tsv',
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers)
else:
valid_pkl_data = load_from_pkl(self.conf.valid_data_path)
valid_data, valid_length, valid_target = valid_pkl_data['data'], valid_pkl_data['length'], valid_pkl_data['target']
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers, chunk_size=self.conf.chunk_size)
if self.conf.test_data_path is not None:
if not self.conf.test_data_path.endswith('.pkl'):
test_data, test_length, test_target = self.problem.encode(self.conf.test_data_path, self.conf.file_columns, self.conf.input_types,
self.conf.file_with_col_header, self.conf.object_inputs, self.conf.answer_column_name, max_lengths=self.conf.max_lengths,
min_sentence_len = self.conf.min_sentence_len, extra_feature = self.conf.extra_feature,fixed_lengths=self.conf.fixed_lengths,
file_format='tsv', show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers)
else:
test_pkl_data = load_from_pkl(self.conf.test_data_path)
test_data, test_length, test_target = test_pkl_data['data'], test_pkl_data['length'], test_pkl_data['target']
test_data, test_length, test_target = self.problem.encode(self.conf.test_data_path, self.conf.file_columns,
self.conf.input_types, self.conf.file_with_col_header, self.conf.object_inputs, self.conf.answer_column_name, max_lengths=self.conf.max_lengths,
min_sentence_len = self.conf.min_sentence_len, extra_feature = self.conf.extra_feature,fixed_lengths=self.conf.fixed_lengths, file_format='tsv',
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers, chunk_size=self.conf.chunk_size)
stop_training = False
epoch = 1
@ -132,26 +120,32 @@ class LearningMachine(object):
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.mrc:
streaming_recoder = StreamingRecorder(['prediction', 'answer_text'])
logging.info("=" * 100 + '\n' + "*" * 15 + 'Start training' + "*" * 15)
while not stop_training and epoch <= self.conf.max_epoch:
logging.info('Training: Epoch ' + str(epoch))
train_data_generator = self._get_training_data_generator()
part_index = 1
for train_data, train_length, train_target in train_data_generator:
logging.debug('Training: Epoch %s Part %s'%(epoch, part_index))
part_index += 1
data_batches, length_batches, target_batches = \
get_batches(self.problem, train_data, train_length, train_target, self.conf.batch_size_total,
self.conf.input_types, None, permutate=True, transform_tensor=True)
whole_batch_num = len(target_batches)
valid_batch_num = max(len(target_batches) // self.conf.valid_times_per_epoch, 1)
if torch.cuda.device_count() > 1:
small_batch_num = whole_batch_num * torch.cuda.device_count() # total batch num over all the gpus
valid_batch_num_show = valid_batch_num * torch.cuda.device_count() # total batch num over all the gpus to do validation
else:
valid_batch_num = min(self.conf.steps_per_validation, whole_batch_num)
small_batch_num = whole_batch_num
valid_batch_num_show = valid_batch_num
batch_num_to_show_results = self.conf.batch_num_to_show_results
if torch.cuda.device_count() > 1:
batch_num_to_show_results *= torch.cuda.device_count() # total batch num overall all the gpus to log
small_batch_num *= torch.cuda.device_count() # total batch num over all the gpus
valid_batch_num_show *= torch.cuda.device_count() # total batch num over all the gpus to do validation
streaming_recoder.clear_records()
all_costs = []
logging.info('There are %d batches during an epoch; validation are conducted every %d batch' % (small_batch_num, valid_batch_num_show))
logging.info('There are %d batches during current period; validation are conducted every %d batch' % (small_batch_num, valid_batch_num_show))
if self.conf.mode == 'normal':
progress = tqdm(range(len(target_batches)))
@ -169,6 +163,8 @@ class LearningMachine(object):
(not self.model.module.layers[tmp_output_layer_id].layer_conf.last_hidden_softmax):
logits_softmax[tmp_output_layer_id] = nn.functional.softmax(
logits[tmp_output_layer_id], dim=-1)
elif isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
pass
else:
logits_softmax[tmp_output_layer_id] = logits[tmp_output_layer_id]
else:
@ -177,6 +173,8 @@ class LearningMachine(object):
(not self.model.layers[tmp_output_layer_id].layer_conf.last_hidden_softmax):
logits_softmax[tmp_output_layer_id] = nn.functional.softmax(
logits[tmp_output_layer_id], dim=-1)
elif isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
pass
else:
logits_softmax[tmp_output_layer_id] = logits[tmp_output_layer_id]
@ -194,6 +192,7 @@ class LearningMachine(object):
prediction_scores_all = None
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging:
logits = list(logits.values())[0]
if not isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
logits_softmax = list(logits_softmax.values())[0]
assert len(logits_softmax.shape) == 3, 'The dimension of your output is %s, but we need [batch_size*GPUs, sequence length, representation dim]' % (str(list(logits_softmax.shape)), )
prediction_scores = None
@ -214,16 +213,25 @@ class LearningMachine(object):
if ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging:
# Transform output shapes for metric evaluation
# for seq_tag_f1 metric
prediction_indices = logits_softmax.data.max(2)[1].cpu().numpy() # [batch_size, seq_len]
if isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
forward_score, scores, masks, tag_seq, transitions, layer_conf = logits
prediction_indices = tag_seq.cpu().numpy()
streaming_recoder.record_one_row([self.problem.decode(prediction_indices, length_batches[i]['target'][self.conf.answer_column_name[0]].numpy()),
prediction_scores, self.problem.decode(target_batches[i][self.conf.answer_column_name[0]],
prediction_scores, self.problem.decode(
target_batches[i][self.conf.answer_column_name[0]],
length_batches[i]['target'][self.conf.answer_column_name[0]].numpy())], keep_dim=False)
else:
prediction_indices = logits_softmax.data.max(2)[1].cpu().numpy() # [batch_size, seq_len]
# pytorch's CrossEntropyLoss only support this
logits_flat[self.conf.output_layer_id[0]] = logits.view(-1, logits.size(2)) # [batch_size * seq_len, # of tags]
#target_batches[i] = target_batches[i].view(-1) # [batch_size * seq_len]
# [batch_size * seq_len]
target_batches[i][self.conf.answer_column_name[0]] = target_batches[i][self.conf.answer_column_name[0]].reshape(-1)
streaming_recoder.record_one_row([self.problem.decode(prediction_indices, length_batches[i]['target'][self.conf.answer_column_name[0]].numpy()),
prediction_scores, self.problem.decode(
target_batches[i][self.conf.answer_column_name[0]],
length_batches[i]['target'][self.conf.answer_column_name[0]].numpy())], keep_dim=False)
target_batches[i][self.conf.answer_column_name[0]] = target_batches[i][
self.conf.answer_column_name[0]].reshape(-1)
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.classification:
prediction_indices = logits_softmax.detach().max(1)[1].cpu().numpy()
@ -260,6 +268,9 @@ class LearningMachine(object):
for single_target in self.conf.answer_column_name:
if isinstance(target_batches[i][single_target], torch.Tensor):
target_batches[i][single_target] = transfer_to_gpu(target_batches[i][single_target])
if isinstance(loss_fn.loss_fn[0], CRFLoss):
loss = loss_fn.loss_fn[0](forward_score, scores, masks, list(target_batches[i].values())[0], transitions, layer_conf)
else:
loss = loss_fn(logits_flat, target_batches[i])
all_costs.append(loss.item())
@ -279,7 +290,7 @@ class LearningMachine(object):
or ProblemTypes[self.problem.problem_type] == ProblemTypes.classification:
del prediction_indices
if show_result_cnt == self.conf.batch_num_to_show_results:
if show_result_cnt == batch_num_to_show_results:
if ProblemTypes[self.problem.problem_type] == ProblemTypes.classification:
result = self.evaluator.evaluate(streaming_recoder.get('target'),
streaming_recoder.get('prediction'), y_pred_pos_score=streaming_recoder.get('pred_scores'),
@ -334,7 +345,7 @@ class LearningMachine(object):
test_data, test_length, test_target = self.problem.encode(test_data_path, self.conf.file_columns, self.conf.input_types,
self.conf.file_with_col_header, self.conf.object_inputs, self.conf.answer_column_name, max_lengths=self.conf.max_lengths,
min_sentence_len = self.conf.min_sentence_len, extra_feature = self.conf.extra_feature,fixed_lengths=self.conf.fixed_lengths, file_format='tsv',
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers)
show_progress=True if self.conf.mode == 'normal' else False, cpu_num_workers=self.conf.cpu_num_workers, chunk_size=self.conf.chunk_size)
else:
test_pkl_data = load_from_pkl(test_data_path)
test_data, test_length, test_target = test_pkl_data['data'], test_pkl_data['length'], test_pkl_data['target']
@ -472,18 +483,29 @@ class LearningMachine(object):
logits_flat = {}
if ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging:
logits = list(logits.values())[0]
if isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
forward_score, scores, masks, tag_seq, transitions, layer_conf = logits
prediction_indices = tag_seq.cpu().numpy()
streaming_recoder.record_one_row(
[self.problem.decode(prediction_indices, length_batches[i]['target'][self.conf.answer_column_name[0]].numpy()),
prediction_pos_scores,
self.problem.decode(target_batches[i], length_batches[i]['target'][self.conf.answer_column_name[0]].numpy())],
keep_dim=False)
else:
logits_softmax = list(logits_softmax.values())[0]
# Transform output shapes for metric evaluation
# for seq_tag_f1 metric
prediction_indices = logits_softmax.data.max(2)[1].cpu().numpy() # [batch_size, seq_len]
streaming_recoder.record_one_row(
[self.problem.decode(prediction_indices, length_batches[i]['target'][self.conf.answer_column_name[0]].numpy()), prediction_pos_scores,
self.problem.decode(target_batches[i], length_batches[i]['target'][self.conf.answer_column_name[0]].numpy())], keep_dim=False)
# pytorch's CrossEntropyLoss only support this
logits_flat[self.conf.output_layer_id[0]] = logits.view(-1, logits.size(2)) # [batch_size * seq_len, # of tags]
#target_batches[i] = target_batches[i].view(-1) # [batch_size * seq_len]
target_batches[i][self.conf.answer_column_name[0]] = target_batches[i][self.conf.answer_column_name[0]].reshape(-1) # [batch_size * seq_len]
streaming_recoder.record_one_row(
[self.problem.decode(prediction_indices, length_batches[i]['target'][self.conf.answer_column_name[0]].numpy()),
prediction_pos_scores,
self.problem.decode(target_batches[i], length_batches[i]['target'][self.conf.answer_column_name[0]].numpy())],
keep_dim=False)
target_batches[i][self.conf.answer_column_name[0]] = target_batches[i][
self.conf.answer_column_name[0]].reshape(-1) # [batch_size * seq_len]
if to_predict:
prediction_batch = self.problem.decode(prediction_indices, length_batches[i][key_random].numpy())
@ -546,8 +568,13 @@ class LearningMachine(object):
predict_stream_recoder.record_one_row([prediction])
if to_predict:
logits_len = len(list(logits.values())[0]) \
if ProblemTypes[self.problem.problem_type] == ProblemTypes.mrc else len(logits)
if ProblemTypes[self.problem.problem_type] == ProblemTypes.mrc:
logits_len = len(list(logits.values())[0])
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging and isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
# for sequence_tagging task, logits is tuple type which index 3 is tag_seq [batch_size*seq_len]
logits_len = logits[3].size(0)
else:
logits_len = len(logits)
for sample_idx in range(logits_len):
while True:
sample = fin.readline().rstrip()
@ -563,6 +590,9 @@ class LearningMachine(object):
for single_target in self.conf.answer_column_name:
if isinstance(target_batches[i][single_target], torch.Tensor):
target_batches[i][single_target] = transfer_to_gpu(target_batches[i][single_target])
if isinstance(loss_fn.loss_fn[0], CRFLoss):
loss = loss_fn.loss_fn[0](forward_score, scores, masks, list(target_batches[i].values())[0], transitions, layer_conf)
else:
loss = loss_fn(logits_flat, target_batches[i])
loss_recoder.record('loss', loss.item())
@ -639,7 +669,7 @@ class LearningMachine(object):
self.conf.file_with_col_header,self.conf.object_inputs, None, min_sentence_len=self.conf.min_sentence_len,
extra_feature=self.conf.extra_feature,max_lengths=self.conf.max_lengths, fixed_lengths=self.conf.fixed_lengths,
file_format='tsv', show_progress=True if self.conf.mode == 'normal' else False,
cpu_num_workers=self.conf.cpu_num_workers)
cpu_num_workers=self.conf.cpu_num_workers, chunk_size=self.conf.chunk_size)
logging.info("Starting predict ...")
self.model.eval()
@ -685,8 +715,13 @@ class LearningMachine(object):
if ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging:
logits = list(logits.values())[0]
if isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
forward_score, scores, masks, tag_seq, transitions, layer_conf = logits
prediction_indices = tag_seq.cpu().numpy()
else:
logits_softmax = list(logits_softmax.values())[0]
# Transform output shapes for metric evaluation
# for seq_tag_f1 metric
prediction_indices = logits_softmax.data.max(2)[1].cpu().numpy() # [batch_size, seq_len]
prediction_batch = self.problem.decode(prediction_indices, length_batches[i][key_random].numpy())
for prediction_sample in prediction_batch:
@ -745,6 +780,107 @@ class LearningMachine(object):
fin.close()
def interactive(self, sample, file_columns, predict_fields=['prediction'], predict_mode='batch'):
""" interactive prediction
Args:
file_columns: representation the columns of sample
predict_mode: interactive|batch(need a predict file)
"""
predict_data, predict_length, _, _, _ = \
self.problem.encode_data_list(sample, file_columns, self.conf.input_types, self.conf.object_inputs, None,
self.conf.min_sentence_len, self.conf.extra_feature, self.conf.max_lengths,
self.conf.fixed_lengths, predict_mode=predict_mode)
if predict_data is None:
return 'Wrong Case!'
self.model.eval()
with torch.no_grad():
data_batches, length_batches, _ = \
get_batches(self.problem, predict_data, predict_length, None, 1,
self.conf.input_types, None, permutate=False, transform_tensor=True, predict_mode=predict_mode)
streaming_recoder = StreamingRecorder(predict_fields)
key_random = random.choice(
list(length_batches[0].keys()).remove('target') if 'target' in list(length_batches[0].keys()) else
list(length_batches[0].keys()))
param_list, inputs_desc, length_desc = transform_params2tensors(data_batches[0], length_batches[0])
logits = self.model(inputs_desc, length_desc, *param_list)
logits_softmax = {}
if isinstance(self.model, nn.DataParallel):
for tmp_output_layer_id in self.model.module.output_layer_id:
if isinstance(self.model.module.layers[tmp_output_layer_id], Linear) and \
(not self.model.module.layers[tmp_output_layer_id].layer_conf.last_hidden_softmax):
logits_softmax[tmp_output_layer_id] = nn.functional.softmax(
logits[tmp_output_layer_id], dim=-1)
else:
logits_softmax[tmp_output_layer_id] = logits[tmp_output_layer_id]
else:
for tmp_output_layer_id in self.model.output_layer_id:
if isinstance(self.model.layers[tmp_output_layer_id], Linear) and \
(not self.model.layers[tmp_output_layer_id].layer_conf.last_hidden_softmax):
logits_softmax[tmp_output_layer_id] = nn.functional.softmax(
logits[tmp_output_layer_id], dim=-1)
else:
logits_softmax[tmp_output_layer_id] = logits[tmp_output_layer_id]
if ProblemTypes[self.problem.problem_type] == ProblemTypes.sequence_tagging:
logits = list(logits.values())[0]
if isinstance(get_layer_class(self.model, tmp_output_layer_id), CRF):
forward_score, scores, masks, tag_seq, transitions, layer_conf = logits
prediction_indices = tag_seq.cpu().numpy()
else:
logits_softmax = list(logits_softmax.values())[0]
# Transform output shapes for metric evaluation
# for seq_tag_f1 metric
prediction_indices = logits_softmax.data.max(2)[1].cpu().numpy() # [batch_size, seq_len]
prediction_batch = self.problem.decode(prediction_indices, length_batches[0][key_random].numpy())
for prediction_sample in prediction_batch:
streaming_recoder.record('prediction', " ".join(prediction_sample))
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.classification:
logits = list(logits.values())[0]
logits_softmax = list(logits_softmax.values())[0]
prediction_indices = logits_softmax.data.max(1)[1].cpu().numpy()
for field in predict_fields:
if field == 'prediction':
streaming_recoder.record(field,
self.problem.decode(prediction_indices,
length_batches[0][key_random].numpy()))
elif field == 'confidence':
prediction_scores = logits_softmax.cpu().data.numpy()
for prediction_score, prediction_idx in zip(prediction_scores, prediction_indices):
streaming_recoder.record(field, prediction_score[prediction_idx])
elif field.startswith('confidence') and field.find('@') != -1:
label_specified = field.split('@')[1]
label_specified_idx = self.problem.output_dict.id(label_specified)
confidence_specified = torch.index_select(logits_softmax.cpu(), 1, torch.tensor([label_specified_idx], dtype=torch.long)).squeeze(1)
streaming_recoder.record(field, confidence_specified.data.numpy())
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.regression:
logits = list(logits.values())[0]
# logits_softmax is unuseful for regression task!
logits_softmax = list(logits_softmax.values())[0]
logits_flat = logits.squeeze(1)
prediction_scores = logits_flat.detach().cpu().numpy()
streaming_recoder.record_one_row([prediction_scores])
elif ProblemTypes[self.problem.problem_type] == ProblemTypes.mrc:
for key, value in logits.items():
logits[key] = value.squeeze()
for key, value in logits_softmax.items():
logits_softmax[key] = value.squeeze()
passage_identify = None
for type_key in data_batches[0].keys():
if 'p' in type_key.lower():
passage_identify = type_key
break
if not passage_identify:
raise Exception('MRC task need passage information.')
prediction = self.problem.decode(logits_softmax, lengths=length_batches[0][passage_identify],
batch_data=data_batches[0][passage_identify])
streaming_recoder.record_one_row([prediction])
return "\t".join([str(streaming_recoder.get(field)[0]) for field in predict_fields])
def load_model(self, model_path):
if self.use_gpu is True:
self.model = torch.load(model_path)
@ -762,5 +898,19 @@ class LearningMachine(object):
logging.info("Model %s loaded!" % model_path)
logging.info("Total trainable parameters: %d" % (get_trainable_param_num(self.model)))
def _get_training_data_generator(self):
if not self.conf.use_cache:
return self.problem.get_encode_generator(self.conf, build_cache=False)
if not self.conf.encoding_file_index:
return self._get_save_encode_generator()
assert self.conf.load_encoding_cache_generator, 'function conf.load_encoding_cache_generator is not defined'
return self.conf.load_encoding_cache_generator(self.conf.encoding_cache_dir, self.conf.encoding_file_index)
def _get_save_encode_generator(self):
load_save_encode_generator = self.problem.get_encode_generator(self.conf, build_cache=True)
for data, lengths, target in load_save_encode_generator:
yield data, lengths, target
cache_index = load_from_json(self.conf.encoding_cache_index_file_path)
self.conf.encoding_file_index = cache_index[st.cencoding_key_index]

Просмотреть файл

@ -18,7 +18,7 @@ EMBED_LAYER_ID = 'embedding'
def get_conf(layer_id, layer_name, input_layer_ids, all_layer_configs, model_input_ids, use_gpu,
conf_dict=None, shared_conf=None, succeed_embedding_flag=False, output_layer_flag=False,
target_num=None, fixed_lengths=None):
target_num=None, fixed_lengths=None, target_dict=None):
""" get layer configuration
Args
@ -51,14 +51,24 @@ def get_conf(layer_id, layer_name, input_layer_ids, all_layer_configs, model_inp
# for classification tasks, we usually add a Linear layer to project the output to dimension of number of classes. If we don't know the #classes, we can use '-1' instead and we would calculate the number of classes from the corpus.
if layer_name == 'Linear':
if isinstance(conf_dict['hidden_dim'], list) and conf_dict['hidden_dim'][-1] == -1:
if isinstance(conf_dict['hidden_dim'], list):
if conf_dict['hidden_dim'][-1] == -1:
assert output_layer_flag is True, "Only in the last layer, hidden_dim == -1 is allowed!"
assert target_num is not None, "Number of targets should be given!"
conf_dict['hidden_dim'][-1] = target_num
elif conf_dict['hidden_dim'][-1] == '#target#':
logging.info('#target# position will be replace by target num: %d' % target_num)
conf_dict['hidden_dim'][-1] = target_num
elif isinstance(conf_dict['hidden_dim'], int) and conf_dict['hidden_dim'] == -1:
assert output_layer_flag is True, "Only in the last layer, hidden_dim == -1 is allowed!"
assert target_num is not None, "Number of targets should be given!"
conf_dict['hidden_dim'] = target_num
elif isinstance(conf_dict['hidden_dim'], str) and conf_dict['hidden_dim'] == '#target#':
logging.info('#target# position will be replace by target num: %d' % target_num)
conf_dict['hidden_dim'] = target_num
# add some necessary attribute for CRF layer
if layer_name == 'CRF':
conf_dict['target_dict'] = target_dict
conf = eval(layer_name + "Conf")(**conf_dict)
except NameError as e:
@ -104,6 +114,8 @@ def get_conf(layer_id, layer_name, input_layer_ids, all_layer_configs, model_inp
# inference and varification inside the layer
conf.inference() # update some attributes which relies on input dimension or something else
conf.verify() # verify if the configuration is legal
former_conf = None if len(all_layer_configs) == 0 else list(all_layer_configs.values())[-1]
conf.verify_former_block(former_conf) # check if has special attribute rely on former layer
logging.debug('Layer id: %s; name: %s; input_dims: %s; input_ranks: %s; output_dim: %s; output_rank: %s' % (layer_id, layer_name, conf.input_dims if layer_id != 'embedding' else 'None', conf.input_ranks, conf.output_dim, conf.output_rank))
@ -211,7 +223,7 @@ class Model(nn.Module):
all_layer_configs[EMBED_LAYER_ID] = get_conf(EMBED_LAYER_ID, layer_arch['layer'],
None, all_layer_configs, inputs, self.use_gpu, conf_dict={'conf': emb_conf},
shared_conf=None, succeed_embedding_flag=False, output_layer_flag=output_layer_flag,
target_num=target_num, fixed_lengths=fixed_lengths_corrected)
target_num=target_num, fixed_lengths=fixed_lengths_corrected, target_dict=problem.output_dict)
self.add_layer(EMBED_LAYER_ID, get_layer(layer_arch['layer'], all_layer_configs[EMBED_LAYER_ID]))
else:
if layer_arch['layer'] in self.layers and not 'conf' in layer_arch:
@ -230,7 +242,7 @@ class Model(nn.Module):
layer_arch['inputs'], all_layer_configs, inputs, self.use_gpu, conf_dict=conf_dict,
shared_conf=shared_conf, succeed_embedding_flag=succeed_embedding_flag,
output_layer_flag=output_layer_flag, target_num=target_num,
fixed_lengths=fixed_lengths_corrected)
fixed_lengths=fixed_lengths_corrected, target_dict=problem.output_dict)
if layer_arch['layer'] in self.layers and not 'conf' in layer_arch:
self.add_layer(layer_arch['layer_id'], self.layers[layer_arch['layer']])
@ -391,7 +403,7 @@ class Model(nn.Module):
return representation_output
def is_cuda(self):
return next(self.parameters()).data.is_cuda
return list(self.parameters())[-1].data.is_cuda
def update_use_gpu(self, new_use_gpu):
self.use_gpu = new_use_gpu

Просмотреть файл

@ -14,8 +14,8 @@ import shutil
from losses.BaseLossConf import BaseLossConf
#import traceback
from settings import LanguageTypes, ProblemTypes, TaggingSchemes, SupportedMetrics, PredictionTypes, DefaultPredictionFields
from utils.common_utils import log_set, prepare_dir
from settings import LanguageTypes, ProblemTypes, TaggingSchemes, SupportedMetrics, PredictionTypes, DefaultPredictionFields, ConstantStatic
from utils.common_utils import log_set, prepare_dir, md5
from utils.exceptions import ConfigurationError
import numpy as np
@ -219,6 +219,10 @@ class ModelConf(object):
# vocabulary setting
self.max_vocabulary = self.get_item(['training_params', 'vocabulary', 'max_vocabulary'], default=800000, use_default=True)
self.min_word_frequency = self.get_item(['training_params', 'vocabulary', 'min_word_frequency'], default=3, use_default=True)
self.max_building_lines = self.get_item(['training_params', 'vocabulary', 'max_building_lines'], default=1000 * 1000, use_default=True)
# chunk_size
self.chunk_size = self.get_item(['training_params', 'chunk_size'], default=1000 * 1000, use_default=True)
# file column header setting
self.file_with_col_header = self.get_item(['inputs', 'file_with_col_header'], default=False, use_default=True)
@ -280,6 +284,9 @@ class ModelConf(object):
tmp_problem_path = os.path.join(self.save_base_dir, '.necessary_cache', 'problem.pkl')
self.problem_path = tmp_problem_path if os.path.isfile(tmp_problem_path) else os.path.join(self.save_base_dir, 'necessary_cache', 'problem.pkl')
# cache configuration
self._load_cache_config_from_conf()
# training params
self.training_params = self.get_item(['training_params'])
@ -303,12 +310,17 @@ class ModelConf(object):
self.max_epoch = self.params.max_epoch
else:
self.max_epoch = self.get_item(['training_params', 'max_epoch'], default=float('inf'))
self.valid_times_per_epoch = self.get_item(['training_params', 'valid_times_per_epoch'], default=1)
if 'valid_times_per_epoch' in self.conf['training_params']:
logging.info("configuration[training_params][valid_times_per_epoch] is deprecated, please use configuration[training_params][steps_per_validation] instead")
self.steps_per_validation = self.get_item(['training_params', 'steps_per_validation'], default=10)
self.batch_num_to_show_results = self.get_item(['training_params', 'batch_num_to_show_results'], default=10)
self.max_lengths = self.get_item(['training_params', 'max_lengths'], default=None, use_default=True)
self.fixed_lengths = self.get_item(['training_params', 'fixed_lengths'], default=None, use_default=True)
if self.fixed_lengths:
self.max_lengths = None
if ProblemTypes[self.problem_type] == ProblemTypes.sequence_tagging:
self.fixed_lengths = None
self.max_lengths = None
if torch.cuda.device_count() > 1:
self.batch_size_total = torch.cuda.device_count() * self.training_params['batch_size']
@ -403,7 +415,8 @@ class ModelConf(object):
"The configuration file %s is illegal. There should be an item configuration[%s], "
"but the item %s is not found." % (self.conf_path, "][".join(error_keys), key))
else:
print("configuration[%s] is not found in %s, use default value %s" % ("][".join(error_keys), self.conf_path, repr(default)))
# print("configuration[%s] is not found in %s, use default value %s" %
# ("][".join(error_keys), self.conf_path, repr(default)))
item = default
return item
@ -525,3 +538,23 @@ class ModelConf(object):
shutil.copy(params.conf_path, self.save_base_dir)
logging.info('Configuration file is backed up to %s' % (self.save_base_dir))
def _load_cache_config_from_conf(self):
# training data
self.train_data_md5 = None
if self.phase == 'train' and self.train_data_path:
logging.info("Calculating the md5 of traing data ...")
self.train_data_md5 = md5([self.train_data_path])
logging.info("the md5 of traing data is %s"%(self.train_data_md5))
# problem
self.problem_md5 = None
# encoding
self.encoding_cache_dir = None
self.encoding_cache_index_file_path = None
self.encoding_cache_index_file_md5_path = None
self.encoding_file_index = None
self.encoding_cache_legal_line_cnt = 0
self.encoding_cache_illegal_line_cnt = 0
self.load_encoding_cache_generator = None

Просмотреть файл

@ -1,4 +1,6 @@
# ***NeuronBlocks*** - Building Your NLP DNN Models Like Playing Lego
<img src="https://i.imgur.com/IanH6xI.png" width="450">
## Building Your NLP DNN Models Like Playing Lego
[![language](https://img.shields.io/badge/language-en%20%7C%20中文-brightgreen.svg)](#language-supported)
[![python](https://img.shields.io/badge/python-3.6%20%7C%203.7-blue.svg)](https://www.python.org)
@ -7,7 +9,7 @@
[简体中文](README_zh_CN.md)
[Tutorial](Tutorial.md) [中文教程](Tutorial_zh_CN.md)
[Tutorial](Tutorial.md) [中文教程](Tutorial_zh_CN.md) [Demo Video](https://youtu.be/x6cOpVSZcdo)
# Table of Contents
@ -29,7 +31,7 @@ NeuronBlocks consists of two major components: ***Block Zoo*** and ***Model Zoo*
- In ***Block Zoo***, we provide commonly used neural network components as building blocks for model architecture design.
- In ***Model Zoo***, we provide a suite of NLP models for common NLP tasks, in the form of **JSON configuration** files.
<img src="https://i.imgur.com/LMD0PFQ.png" width="300">
<img src="https://i.imgur.com/LMD0PFQ.png" width="250">
### <span id="language-supported">Language Supported</span>
- English
@ -50,6 +52,9 @@ Users can either pick existing models (config files) in *Model Zoo* to start mod
<img src="https://i.imgur.com/q0p6Wvz.png" width="300">
<img src="https://i.imgur.com/lFaBtnh.png" width="700">
# Get Started in 60 Seconds
## <span id="installation">Installation</span>
@ -93,6 +98,21 @@ python test.py --conf_path=model_zoo/demo/conf.json
# predict
python predict.py --conf_path=model_zoo/demo/conf.json
```
For prediction, NeuronBlocks have two modes: **Interactive** and **Batch**.
- *Interactive Prediction Mode:* The interactive mode provides interactive interface, users can input case according to corresponding prompt message and get realtime prediction result from trained model, and input "exit" to exit interactive interface.
```bash
# use the above example
# interactive prediction
python predict.py --conf_path=model_zoo/demo/conf.json --predict_mode='interactive'
```
- *Batch Prediction Mode:* For batched cases prediction, NeuronBlocks provides batch prediction mode which receives a cases file as input and write the prediction results in the prediction file.
```bash
# use the above example
# batch prediction
python predict.py --conf_path=model_zoo/demo/conf.json --predict_mode='batch' --predict_data_path=dataset/demo/predict.tsv
```
For more details, please refer to [Tutorial.md](Tutorial.md) and [Code documentation](https://microsoft.github.io/NeuronBlocks/).
# Who should consider using NeuronBlocks
@ -134,7 +154,7 @@ Anyone who are familiar with are highly encouraged to contribute code.
```
@article{gong2019neuronblocks,
title={NeuronBlocks--Building Your NLP DNN Models Like Playing Lego},
author={Gong, Ming and Shou, Linjun and Lin, Wutao and Sang, Zhijie and Yan, Quanjia and Yang, Ze and Jiang, Daxin},
author={Gong, Ming and Shou, Linjun and Lin, Wutao and Sang, Zhijie and Yan, Quanjia and Yang, Ze, Cheng, Feixiang and Jiang, Daxin},
journal={arXiv preprint arXiv:1904.09535},
year={2019}
}
@ -155,5 +175,5 @@ If you have any questions, please contact NeuronBlocks@microsoft.com
If you have wechat, you can also add the following account:
<img src="https://i.imgur.com/lI2oQWo.jpg" width="200">
<img src="https://i.imgur.com/UfOYvt1.jpg" width="200">

Просмотреть файл

@ -1,4 +1,6 @@
# ***NeuronBlocks*** - 像搭积木一样构建自然语言理解深度学习模型
<img src="https://i.imgur.com/IanH6xI.png" width="450">
## 像搭积木一样构建自然语言理解深度学习模型
[![language](https://img.shields.io/badge/language-en%20%7C%20中文-brightgreen.svg)](#language-supported)
[![python](https://img.shields.io/badge/python-3.6%20%7C%203.7-blue.svg)](https://www.python.org)
@ -7,7 +9,7 @@
[English version](README.md)
[中文教程](Tutorial_zh_CN.md) [Tutorial](Tutorial.md)
[中文教程](Tutorial_zh_CN.md) [Tutorial](Tutorial.md) [Demo Video](https://youtu.be/x6cOpVSZcdo)
# 目录
@ -47,11 +49,12 @@ NeuronBlocks包括 ***Block Zoo*** 和 ***Model Zoo*** 两个重要组件,其
- 更多……
### 使用方法
用户可以选择 *Model Zoo* 中的示例模型JSON配置文件开启模型训练或者利用 *Block Zoo* 中的神经网络模块构建新的模型,就像玩乐高积木一样。
<img src="https://i.imgur.com/q0p6Wvz.png" width="300">
<img src="https://i.imgur.com/lFaBtnh.png" width="700">
# 快速入门
## 安装
@ -95,6 +98,19 @@ python test.py --conf_path=model_zoo/demo/conf.json
python predict.py --conf_path=model_zoo/demo/conf.json
```
对于预测NeuronBlocks 提供了两种预测的形式: **交互式**和**批量式**
- *交互式:* 交互式模式预测提供了交互界面,用户可以根据输入提示信息每次输入一个样本并实时得到模型前向计算出的结果,输入 "exit" 时退出交互预测模式。
```bash
# use the above example
# interactive prediction
python predict.py --conf_path=model_zoo/demo/conf.json --predict_mode='interactive'
```
- *批量式:* 对于批量样本预测的需求NeuronBlocks 提供批量预测模式,其接受一个包含批量样本的文件作为输入,并且将模型前向计算的结果写回这个文件。
```bash
# use the above example
# batch prediction
python predict.py --conf_path=model_zoo/demo/conf.json --predict_mode='batch' --predict_data_path=dataset/demo/predict.tsv
```
更多细节, 请查看[Tutorial_zh_CN.md](Tutorial_zh_CN.md) 和 [Code documentation](https://microsoft.github.io/NeuronBlocks/)。
# 适用人群
@ -135,7 +151,7 @@ NeuronBlocks以开放的模式运行。它由 **微软 STCA NLP Group** 设计
```
@article{gong2019neuronblocks,
title={NeuronBlocks--Building Your NLP DNN Models Like Playing Lego},
author={Gong, Ming and Shou, Linjun and Lin, Wutao and Sang, Zhijie and Yan, Quanjia and Yang, Ze and Jiang, Daxin},
author={Gong, Ming and Shou, Linjun and Lin, Wutao and Sang, Zhijie and Yan, Quanjia and Yang, Ze, Cheng, Feixiang and Jiang, Daxin},
journal={arXiv preprint arXiv:1904.09535},
year={2019}
}
@ -156,4 +172,5 @@ Licensed under the [MIT](LICENSE) License.
如果您有微信,也可以添加工具包的官方账号:
<img src="https://i.imgur.com/lI2oQWo.jpg" width="200">
<img src="https://i.imgur.com/UfOYvt1.jpg" width="200">

Просмотреть файл

@ -1,4 +1,4 @@
# ***NeuronBlocks*** Tutorial
# <img src="https://i.imgur.com/YLrkvW3.png" width="80"> ***NeuronBlocks*** Tutorial
[简体中文](Tutorial_zh_CN.md)
@ -21,6 +21,7 @@
4. [Compression for MRC Model](#task-6.4)
* [Task 7: Chinese Sentiment Analysis](#task-7)
* [Task 8: Chinese Text Matching](#task-8)
* [Task 9: Sequence Labeling](#task-9)
* [Advanced Usage](#advanced-usage)
* [Extra Feature Support](#extra-feature)
* [Learning Rate Decay](#lr-decay)
@ -146,10 +147,12 @@ The architecture of the configuration file is:
CUDA_VISIBLE_DEVICES= python train.py
```
- ***cpu_num_workers***. [default: -1] Define the number of processes to preprocess the dataset. The number of processes is equal to that of logical cores CPU supports if value is negtive or 0, otherwise it is equal to *cpu_num_workers*.
- ***chunk_size***. [default: 1000000] Define the chunk size of files that NB reads every time for avoiding out of memory and the mechanism of lazy-loading.
- ***batch_size***. Define the batch size here. If there are multiple GPUs, *batch_size* is the batch size of each GPU.
- ***batch_num_to_show_results***. [necessary for training] During the training process, show the results every batch_num_to_show_results batches.
- ***max_epoch***. [necessary for training] The maximum number of epochs to train.
- ***valid_times_per_epoch***. [optional for training, default: 1] Define how many times to conduct validation per epoch. Usually, we conduct validation after each epoch, but for a very large corpus, we'd better validate multiple times in case to miss the best state of our model. The default value is 1.
- ~~***valid_times_per_epoch***~~. [**deprecated**] Please use steps_per_validation instead.
- ***steps_per_validation***. [default: 10] Define how many steps does each validation take place.
- ***tokenizer***. [optional] Define tokenizer here. Currently, we support 'nltk' and 'jieba'. By default, 'nltk' for English and 'jieba' for Chinese.
- **architecture**. Define the model architecture. The node is a list of layers (blocks) in block_zoo to represent a model. The supported layers of this toolkit are given in [block_zoo overview](https://microsoft.github.io/NeuronBlocks).
@ -299,6 +302,8 @@ Question answer matching is a crucial subtask of the question answering problem,
CNN (NeuronBlocks) | 0.747
BiLSTM (NeuronBlocks) | 0.767
BiLSTM+Attn (NeuronBlocks) | 0.754
[ARC-I](https://arxiv.org/abs/1503.03244) (NeuronBlocks) | 0.7508
[ARC-II](https://arxiv.org/abs/1503.03244) (NeuronBlocks) | 0.7612
[MatchPyramid](https://arxiv.org/abs/1602.06359) (NeuronBlocks) | 0.763
BiLSTM+Match Attention (NeuronBlocks) | 0.786
@ -457,7 +462,7 @@ This task is to train a query regression model to learn from a heavy teacher mod
3. Calculate AUC metric
```bash
cd PROJECT_ROOT
python tools/calculate_AUC.py --input_file models/kdqbc_bilstmattn_cnn/train/predict.tsv --predict_index 2 --label_index 1
python tools/calculate_auc.py --input_file models/kdqbc_bilstmattn_cnn/train/predict.tsv --predict_index 2 --label_index 1
```
*Tips: you can try different models by running different JSON config files.*
@ -501,7 +506,7 @@ This task is to train a query-passage regression model to learn from a heavy tea
3. Calculate AUC metric
```bash
cd PROJECT_ROOT
python tools/calculate_AUC.py --input_file=models/kdtm_match_linearAttn/predict.tsv --predict_index=3 --label_index=2
python tools/calculate_auc.py --input_file=models/kdtm_match_linearAttn/predict.tsv --predict_index=3 --label_index=2
```
*Tips: you can try different models by running different JSON config files.*
@ -562,6 +567,57 @@ Here is an example using Chinese data, for text matching task.
```
*Tips: you can try different models by running different JSON config files. The model file and train log file can be found in JOSN config file's outputs/save_base_dir after you finish training.*
### <span id="task-9">Task 9: Sequence Labeling</span>
Sequence Labeling is an important NLP task, which includes NER, Slot Tagging, Pos Tagging, etc.
- ***Dataset***
[CoNLL 2003](https://www.clips.uantwerpen.be/conll2003/) is a popular dataset in Sequence Labeling task. We use CoNLL 2003 English NER data for our experiment and you can refer the data format in [sample data](https://github.com/microsoft/NeuronBlocks/tree/master/dataset/slot_tagging/conll_2003).
- ***Tagging Scheme***
- NeuronBlocks support both BIO and BIOES tag schemes.
- The IOB scheme is not supported, because of its worse performance in most [experiment](https://arxiv.org/pdf/1707.06799.pdf).
- NeuronBlocks provides a [script](tools/tagging_schemes_converter.py) that converts the tag scheme among IOB/BIO/BIOES (NOTE: the script only supports tsv file which has data and label in two columns).
- ***Usages***
1. Softmax output.
```bash
# train model
cd PROJECT_ROOT
python train.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json
# test model
cd PROJECT_ROOT
python test.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json
```
2. CRF output.
```bash
# train model
cd PROJECT_ROOT
python train.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json
# test model
cd PROJECT_ROOT
python test.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json
```
*Tips: you can try more model in [here](https://github.com/microsoft/NeuronBlocks/tree/master/model_zoo/nlp_tasks/slot_tagging).*
- ***Result***
The result on CoNLL 2003 English NER dataset.
Model | F1-score
-------- | --------
[Ma and Hovy(2016)](https://arxiv.org/pdf/1603.01354.pdf)|87.00
[BiLSTM+Softmax](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json) (NeuronBlocks)|88.50
[Lample et al.(2016)](https://arxiv.org/pdf/1603.01360.pdf)| 89.15
[CLSTM+WLSTM+CRF](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_clstm_wlstm_crf.json) (NeuronBlocks)|90.83
[Chiu and Nichols(2016)](https://www.mitpressjournals.org/doi/pdf/10.1162/tacl_a_00104)|90.91
[CCNN+WLSTM+CRF](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json) (NeuronBlocks)|91.38
*Tips: C means Char and W means Word. CCNN means Char-level representation with CNN model and CLSTM means Char-level representation with LSTM model.*
## <span id="advanced-usage">Advanced Usage</span>
@ -675,5 +731,7 @@ To solve the above problems, NeuronBlocks supports *fixing embedding weight* (em
***training_params/vocabulary/max_vocabulary***. [int, optional for training, default: 800,000] The max size of corpus vocabulary. If corpus vocabulary size is larger than *max_vocabulary*, it will be cut according to word frequency.
***training_params/vocabulary/max_building_lines***. [int, optional for training, default: 1,000,000] The max lines NB will read from every file to build vocabulary
## <span id="faq">Frequently Asked Questions</span>

Просмотреть файл

@ -1,4 +1,4 @@
# ***NeuronBlocks*** 教程
# <img src="https://i.imgur.com/YLrkvW3.png" width="80"> ***NeuronBlocks*** 教程
[English Version](Tutorial.md)
@ -21,6 +21,7 @@
4. [机器阅读理解模型的模型压缩](#task-6.4)
* [任务 7: 中文情感分析](#task-7)
* [任务 8中文文本匹配](#task-8)
* [任务 9序列标注](#task-9)
* [高阶用法](#advanced-usage)
* [额外的feature](#extra-feature)
* [学习率衰减](#lr-decay)
@ -136,10 +137,12 @@ python predict.py --conf_path=model_zoo/demo/conf.json
CUDA_VISIBLE_DEVICES= python train.py
```
- ***cpu_num_workers***. [default: -1] Define the number of processes to preprocess the dataset. The number of processes is equal to that of logical cores CPU supports if value is negtive or 0, otherwise it is equal to *cpu_num_workers*.
- ***chunk_size***. [default: 1000000] Define the chunk size of files that NB reads every time for avoiding out of memory and the mechanism of lazy-loading.
- ***batch_size***. Define the batch size here. If there are multiple GPUs, *batch_size* is the batch size of each GPU.
- ***batch_num_to_show_results***. [necessary for training] During the training process, show the results every batch_num_to_show_results batches.
- ***max_epoch***. [necessary for training] The maximum number of epochs to train.
- ***valid_times_per_epoch***. [optional for training, default: 1] Define how many times to conduct validation per epoch. Usually, we conduct validation after each epoch, but for a very large corpus, we'd better validate multiple times in case to miss the best state of our model. The default value is 1.
- ~~***valid_times_per_epoch***~~. [**deprecated**] Please use steps_per_validation instead.
- ***steps_per_validation***. [default: 10] Define how many steps does each validation take place.
- ***tokenizer***. [optional] Define tokenizer here. Currently, we support 'nltk' and 'jieba'. By default, 'nltk' for English and 'jieba' for Chinese.
- **architecture**. Define the model architecture. The node is a list of layers (blocks) in block_zoo to represent a model. The supported layers of this toolkit are given in [block_zoo overview](https://microsoft.github.io/NeuronBlocks).
@ -288,6 +291,8 @@ Question answer matching is a crucial subtask of the question answering problem,
CNN (NeuronBlocks) | 0.747
BiLSTM (NeuronBlocks) | 0.767
BiLSTM+Attn (NeuronBlocks) | 0.754
[ARC-I](https://arxiv.org/abs/1503.03244) (NeuronBlocks) | 0.7508
[ARC-II](https://arxiv.org/abs/1503.03244) (NeuronBlocks) | 0.7612
[MatchPyramid](https://arxiv.org/abs/1602.06359) (NeuronBlocks) | 0.763
BiLSTM+Match Attention (NeuronBlocks) | 0.786
@ -446,7 +451,7 @@ This task is to train a query regression model to learn from a heavy teacher mod
3. Calculate AUC metric
```bash
cd PROJECT_ROOT
python tools/calculate_AUC.py --input_file models/kdqbc_bilstmattn_cnn/train/predict.tsv --predict_index 2 --label_index 1
python tools/calculate_auc.py --input_file models/kdqbc_bilstmattn_cnn/train/predict.tsv --predict_index 2 --label_index 1
```
*Tips: you can try different models by running different JSON config files.*
@ -490,7 +495,7 @@ This task is to train a query-passage regression model to learn from a heavy tea
3. Calculate AUC metric
```bash
cd PROJECT_ROOT
python tools/calculate_AUC.py --input_file=models/kdtm_match_linearAttn/predict.tsv --predict_index=3 --label_index=2
python tools/calculate_auc.py --input_file=models/kdtm_match_linearAttn/predict.tsv --predict_index=3 --label_index=2
```
*Tips: you can try different models by running different JSON config files.*
@ -552,6 +557,58 @@ This task is to train a query-passage regression model to learn from a heavy tea
```
*提示您可以通过运行不同的JSON配置文件来尝试不同的模型。当训练完成后模型文件和训练日志文件可以在JSON配置的outputs/save_base_dir目录中找到。*
### <span id="task-9">任务 9: 序列标注</span>
序列标注是一项重要的NLP任务包括 NER, Slot Tagging, Pos Tagging 等任务。
- ***数据集***
在序列标注任务中,[CoNLL 2003](https://www.clips.uantwerpen.be/conll2003/)是一个很常用的数据集。在我们的序列标注任务中,使用 CoNLL 2003 中英文 NER 数据作为实验数据,其中数据格式可以参考我们给出的[抽样数据](https://github.com/microsoft/NeuronBlocks/tree/master/dataset/slot_tagging/conll_2003)。
- ***标注策略***
- NeuronBlocks 支持 BIO 和 BIOES 标注策略。
- IOB 标注标注是不被支持的,因为在大多[实验](https://arxiv.org/pdf/1707.06799.pdf)中它具有很差的表现。
- NeuronBlocks 提供一个在不同标注策略(IOB/BIO/BIOES)中的[转化脚本](tools/tagging_schemes_converter.py)(脚本仅支持具有 数据和标签 的两列tsv文件输入)。
- ***用法***
1. Softmax 输出.
```bash
# train model
cd PROJECT_ROOT
python train.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json
# test model
cd PROJECT_ROOT
python test.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json
```
2. CRF 输出.
```bash
# train model
cd PROJECT_ROOT
python train.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json
# test model
cd PROJECT_ROOT
python test.py --conf_path=model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json
```
*提示 :尝试更多模型可 [点击](https://github.com/microsoft/NeuronBlocks/tree/master/model_zoo/nlp_tasks/slot_tagging)。*
- ***结果***
实验采用 CoNLL 2003 英文 NER 数据集。
Model | F1-score
-------- | --------
[Ma and Hovy(2016)](https://arxiv.org/pdf/1603.01354.pdf)|87.00
[BiLSTM+Softmax](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging.json) (NeuronBlocks)|88.50
[Lample et al.(2016)](https://arxiv.org/pdf/1603.01360.pdf)| 89.15
[CLSTM+WLSTM+CRF](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_clstm_wlstm_crf.json) (NeuronBlocks)|90.83
[Chiu and Nichols(2016)](https://www.mitpressjournals.org/doi/pdf/10.1162/tacl_a_00104)|90.91
[CCNN+WLSTM+CRF](https://github.com/microsoft/NeuronBlocks/blob/master/model_zoo/nlp_tasks/slot_tagging/conf_slot_tagging_ccnn_wlstm_crf.json) (NeuronBlocks)|91.38
*提示 : C 代表字符W 代表单词。 CCNN 代表使用 CNN 模型的字符级别表示, CLSTM 代表使用 LSTM 模型的字符级别表示。*
## <span id="advanced-usage">高阶用法</span>
After building a model, the next goal is to train a model with good performance. It depends on a highly expressive model and tricks of the model training. NeuronBlocks provides some tricks of model training.
@ -664,4 +721,6 @@ To solve the above problems, NeuronBlocks supports *fixing embedding weight* (em
***training_params/vocabulary/max_vocabulary***. [int, optional for training, default: 800,000] The max size of corpus vocabulary. If corpus vocabulary size is larger than *max_vocabulary*, it will be cut according to word frequency.
***training_params/vocabulary/max_building_lines***. [int, optional for training, default: 1,000,000] The max lines NB will read from every file to build vocabulary
## <span id="faq">常见问题与答案</span>

0
autotest.sh Normal file → Executable file
Просмотреть файл

Просмотреть файл

@ -1,6 +1,3 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import re
from calculate_AUC import main
@ -10,24 +7,29 @@ task_dir = ['/20_newsgroup_bilstm_attn', '/chinese_text_matching', '/question_pa
results = {'english_text_matching': [0.96655], 'chinese_text_matching': [0.70001], 'quora_question_pairs': [0.72596], 'knowledge_distillation': [0.66329]}
for each_dir, key in zip(task_dir, results.keys()):
target_dir = base_dir + each_dir
try:
with open(target_dir + '/train_autotest.log', 'r') as f_r:
last_line = f_r.readlines()[-1].strip()
score = ''.join(re.findall(r'(?<=accuracy:).*?(?=loss|;)', last_line))
try:
results[key].append(float(score))
except:
results[key].append('wrong')
results[key].append('wrong number in train log')
print ('GPU test. Wrong number in %s/train_autotest.log' %target_dir)
except:
results[key].append('no train log')
try:
with open(target_dir + '/test_autotest.log', 'r') as f_r:
last_line = f_r.readlines()[-1].strip()
score = ''.join(re.findall(r'(?<=accuracy:).*?(?=loss|;)', last_line))
try:
results[key].append(float(score))
except:
results[key].append('wrong')
results[key].append('wrong number in test log')
print ('CPU test. Wrong number in %s/test_autotest.log' %target_dir)
except:
results[key].append('no test log')
# for kdtm_match_linearAttn task, we use calculate_AUC.main()
params = {'input_file': './autotest/models/kdtm_match_linearAttn/predict.tsv', 'predict_index': '3', 'label_index': '2', 'header': False}

Просмотреть файл

@ -154,6 +154,11 @@ class BaseConf(ABC):
# To check if deepcopy is applied
assert id(self.output_dim) != id(self.input_dims[0]), 'Please use copy.deepcopy to copy the input_dim to output_dim'
def verify_former_block(self, former_conf):
"""check if has special attribute rely on former layer
"""
return True
def add_attr_type_assertion(self, attr, specified_type):
""" check if the types of attributes are legal

244
block_zoo/CRF.py Normal file
Просмотреть файл

@ -0,0 +1,244 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
from block_zoo.BaseLayer import BaseLayer, BaseConf
from utils.DocInherit import DocInherit
import torch
import torch.nn as nn
from copy import deepcopy
import torch.autograd as autograd
def argmax(vec):
# return the argmax as a python int
_, idx = torch.max(vec, 1)
return idx.item()
def log_sum_exp(vec, m_size):
"""
calculate log of exp sum
args:
vec (batch_size, vanishing_dim, hidden_dim) : input tensor
m_size : hidden_dim
return:
batch_size, hidden_dim
"""
_, idx = torch.max(vec, 1) # B * 1 * M
max_score = torch.gather(vec, 1, idx.view(-1, 1, m_size)).view(-1, 1, m_size) # B * M
return max_score.view(-1, m_size) + torch.log(torch.sum(torch.exp(vec - max_score.expand_as(vec)), 1)).view(-1, m_size) # B * M
class CRFConf(BaseConf):
"""
Configuration of CRF layer
Args:
"""
def __init__(self, **kwargs):
super(CRFConf, self).__init__(**kwargs)
@DocInherit
def default(self):
self.START_TAG = "<start>"
self.STOP_TAG = "<eos>"
@DocInherit
def declare(self):
self.num_of_inputs = 1
self.input_ranks = [3]
@DocInherit
def inference(self):
self.output_dim = [1]
# add target dict judgement start or end
self.target_dict = deepcopy(self.target_dict.cell_id_map)
if not self.target_dict.get(self.START_TAG):
self.target_dict[self.START_TAG] = len(self.target_dict)
if not self.target_dict.get(self.STOP_TAG):
self.target_dict[self.STOP_TAG] = len(self.target_dict)
super(CRFConf, self).inference()
@DocInherit
def verify(self):
super(CRFConf, self).verify()
class CRF(BaseLayer):
""" Conditional Random Field layer
Args:
layer_conf(CRFConf): configuration of CRF layer
"""
def __init__(self, layer_conf):
super(CRF, self).__init__(layer_conf)
self.target_size = len(self.layer_conf.target_dict)
init_transitions = torch.zeros(self.target_size, self.target_size)
init_transitions[:, self.layer_conf.target_dict[self.layer_conf.START_TAG]] = -10000.0
init_transitions[self.layer_conf.target_dict[self.layer_conf.STOP_TAG], :] = -10000.0
init_transitions[:, 0] = -10000.0
init_transitions[0, :] = -10000.0
if self.layer_conf.use_gpu:
init_transitions = init_transitions.cuda()
self.transitions = nn.Parameter(init_transitions)
def _calculate_forward(self, feats, mask):
"""
input:
feats: (batch, seq_len, self.tag_size)
masks: (batch, seq_len)
"""
batch_size = feats.size(0)
seq_len = feats.size(1)
tag_size = feats.size(2)
mask = mask.transpose(1, 0).contiguous()
ins_num = seq_len * batch_size
# be careful the view shape, it is .view(ins_num, 1, tag_size) but not .view(ins_num, tag_size, 1)
feats = feats.transpose(1, 0).contiguous().view(ins_num, 1, tag_size).expand(ins_num, tag_size, tag_size)
# need to consider start
scores = feats + self.transitions.view(1, tag_size, tag_size).expand(ins_num, tag_size, tag_size)
scores = scores.view(seq_len, batch_size, tag_size, tag_size)
# build iter
seq_iter = enumerate(scores)
_, inivalues = next(seq_iter) # bat_size * from_target_size * to_target_size
# only need start from start_tag
partition = inivalues[:, self.layer_conf.target_dict[self.layer_conf.START_TAG], :].clone().view(batch_size, tag_size, 1) # bat_size * to_target_size
for idx, cur_values in seq_iter:
# previous to_target is current from_target
# partition: previous results log(exp(from_target)), #(batch_size * from_target)
# cur_values: bat_size * from_target * to_target
cur_values = cur_values + partition.contiguous().view(batch_size, tag_size, 1).expand(batch_size, tag_size, tag_size)
cur_partition = log_sum_exp(cur_values, tag_size)
# (bat_size * from_target * to_target) -> (bat_size * to_target)
# partition = utils.switch(partition, cur_partition, mask[idx].view(bat_size, 1).expand(bat_size, self.tagset_size)).view(bat_size, -1)
mask_idx = mask[idx, :].view(batch_size, 1).expand(batch_size, tag_size)
# effective updated partition part, only keep the partition value of mask value = 1
masked_cur_partition = cur_partition.masked_select(mask_idx)
# let mask_idx broadcastable, to disable warning
mask_idx = mask_idx.contiguous().view(batch_size, tag_size, 1)
# replace the partition where the maskvalue=1, other partition value keeps the same
partition.masked_scatter_(mask_idx, masked_cur_partition)
# until the last state, add transition score for all partition (and do log_sum_exp) then select the value in STOP_TAG
cur_values = self.transitions.view(1, tag_size, tag_size).expand(batch_size, tag_size, tag_size) + partition.contiguous().view(batch_size, tag_size, 1).expand(batch_size, tag_size, tag_size)
cur_partition = log_sum_exp(cur_values, tag_size)
final_partition = cur_partition[:, self.layer_conf.target_dict[self.layer_conf.STOP_TAG]]
return final_partition.sum(), scores
def _viterbi_decode(self, feats, mask):
"""
input:
feats: (batch, seq_len, self.tag_size)
mask: (batch, seq_len)
output:
decode_idx: (batch, seq_len) decoded sequence
path_score: (batch, 1) corresponding score for each sequence
"""
batch_size = feats.size(0)
seq_len = feats.size(1)
tag_size = feats.size(2)
# calculate sentence length for each sentence
length_mask = torch.sum(mask.long(), dim=1).view(batch_size, 1).long()
# mask to (seq_len, batch_size)
mask = mask.transpose(1, 0).contiguous()
ins_num = seq_len * batch_size
# be careful the view shape, it is .view(ins_num, 1, tag_size) but not .view(ins_num, tag_size, 1)
feats = feats.transpose(1, 0).contiguous().view(ins_num, 1, tag_size).expand(ins_num, tag_size, tag_size)
# need to consider start
scores = feats + self.transitions.view(1, tag_size, tag_size).expand(ins_num, tag_size, tag_size)
scores = scores.view(seq_len, batch_size, tag_size, tag_size)
# build iter
seq_iter = enumerate(scores)
# record the position of best score
back_points = list()
partition_history = list()
# reverse mask (bug for mask = 1- mask, use this as alternative choice)
mask = (1 - mask.long()).byte()
_, inivalues = next(seq_iter) # bat_size * from_target_size * to_target_size
# only need start from start_tag
partition = inivalues[:, self.layer_conf.target_dict[self.layer_conf.START_TAG], :].clone().view(batch_size, tag_size) # bat_size * to_target_size
# print "init part:",partition.size()
partition_history.append(partition)
# iter over last scores
for idx, cur_values in seq_iter:
# previous to_target is current from_target
# partition: previous results log(exp(from_target)), #(batch_size * from_target)
# cur_values: batch_size * from_target * to_target
cur_values = cur_values + partition.contiguous().view(batch_size, tag_size, 1).expand(batch_size, tag_size, tag_size)
partition, cur_bp = torch.max(cur_values, 1)
partition_history.append(partition)
# cur_bp: (batch_size, tag_size) max source score position in current tag
# set padded label as 0, which will be filtered in post processing
cur_bp.masked_fill_(mask[idx].view(batch_size, 1).expand(batch_size, tag_size), 0)
back_points.append(cur_bp)
# add score to final STOP_TAG
partition_history = torch.cat(partition_history, 0).view(seq_len, batch_size, -1).transpose(1, 0).contiguous() # (batch_size, seq_len. tag_size)
# get the last position for each setences, and select the last partitions using gather()
last_position = length_mask.view(batch_size, 1, 1).expand(batch_size, 1, tag_size) - 1
last_partition = torch.gather(partition_history, 1, last_position).view(batch_size,tag_size,1)
# calculate the score from last partition to end state (and then select the STOP_TAG from it)
last_values = last_partition.expand(batch_size, tag_size, tag_size) + self.transitions.view(1, tag_size, tag_size).expand(batch_size, tag_size, tag_size)
_, last_bp = torch.max(last_values, 1)
pad_zero = autograd.Variable(torch.zeros(batch_size, tag_size)).long()
if self.layer_conf.use_gpu:
pad_zero = pad_zero.cuda()
back_points.append(pad_zero)
back_points = torch.cat(back_points).view(seq_len, batch_size, tag_size)
# select end ids in STOP_TAG
pointer = last_bp[:, self.layer_conf.target_dict[self.layer_conf.STOP_TAG]]
insert_last = pointer.contiguous().view(batch_size, 1, 1).expand(batch_size, 1, tag_size)
back_points = back_points.transpose(1, 0).contiguous()
# move the end ids(expand to tag_size) to the corresponding position of back_points to replace the 0 values
back_points.scatter_(1, last_position, insert_last)
back_points = back_points.transpose(1, 0).contiguous()
# decode from the end, padded position ids are 0, which will be filtered if following evaluation
decode_idx = autograd.Variable(torch.LongTensor(seq_len, batch_size))
if self.layer_conf.use_gpu:
decode_idx = decode_idx.cuda()
decode_idx[-1] = pointer.detach()
for idx in range(len(back_points)-2, -1, -1):
pointer = torch.gather(back_points[idx], 1, pointer.contiguous().view(batch_size, 1))
decode_idx[idx] = pointer.detach().view(batch_size)
path_score = None
decode_idx = decode_idx.transpose(1, 0)
return path_score, decode_idx
def forward(self, string, string_len):
"""
CRF layer process: include use transition matrix compute score and viterbi decode
Args:
string(Tensor): [batch_size, seq_len, target_num]
string_len(Tensor): [batch_size]
Returns:
score: the score by CRF inference
best_path: the best bath of viterbi decode
"""
assert string_len is not None, "CRF layer need string length for mask."
masks = []
string_len_val = string_len.cpu().data.numpy()
for i in range(len(string_len)):
masks.append(
torch.cat([torch.ones(string_len_val[i]), torch.zeros(string.shape[1] - string_len_val[i])]))
masks = torch.stack(masks).view(string.shape[0], string.shape[1]).byte()
if self.layer_conf.use_gpu:
masks = masks.cuda()
forward_score, scores = self._calculate_forward(string, masks)
_, tag_seq = self._viterbi_decode(string, masks)
return (forward_score, scores, masks, tag_seq, self.transitions, self.layer_conf), string_len

Просмотреть файл

@ -35,6 +35,9 @@ class ConvConf(BaseConf):
self.output_channel_num = 16
self.batch_norm = True
self.activation = 'ReLU'
self.padding_type = 'VALID'
self.dropout = 0
self.remind_lengths = True
@DocInherit
def declare(self):
@ -43,8 +46,15 @@ class ConvConf(BaseConf):
@DocInherit
def inference(self):
if self.padding_type == 'SAME':
self.padding = int((self.window_size-1)/2)
self.output_dim = [-1]
if self.input_dims[0][1] != -1:
if self.padding_type == 'SAME':
self.output_dim.append(self.input_dims[0][1])
else:
self.output_dim.append((self.input_dims[0][1] - self.window_size) // self.stride + 1)
else:
self.output_dim.append(-1)
@ -67,6 +77,13 @@ class ConvConf(BaseConf):
for attr in necessary_attrs_for_user:
self.add_attr_exist_assertion_for_user(attr)
@DocInherit
def verify_former_block(self, former_conf):
if 'conv' in str(type(former_conf)).lower():
self.mask = False
else:
self.mask = True
class Conv(BaseLayer):
""" Convolution along just 1 direction
@ -82,16 +99,29 @@ class Conv(BaseLayer):
else:
self.activation = None
self.filters = nn.ParameterList([nn.Parameter(torch.randn(layer_conf.output_channel_num,
layer_conf.input_channel_num, layer_conf.window_size, layer_conf.input_dims[0][-1],
requires_grad=True).float())])
self.conv = nn.Conv1d(layer_conf.input_dims[0][-1], layer_conf.output_channel_num, kernel_size=layer_conf.window_size, padding=layer_conf.padding)
if layer_conf.batch_norm:
self.batch_norm = nn.BatchNorm2d(layer_conf.output_channel_num) # the output_chanel of Conv is the input_channel of BN
# self.batch_norm = nn.BatchNorm2d(layer_conf.output_channel_num) # the output_chanel of Conv is the input_channel of BN
self.batch_norm = nn.BatchNorm1d(layer_conf.output_channel_num)
else:
self.batch_norm = None
def forward(self, string, string_len=None):
if layer_conf.dropout > 0:
self.cov_dropout = nn.Dropout(layer_conf.dropout)
else:
self.cov_dropout = None
if layer_conf.use_gpu:
self.conv = self.conv.cuda()
if self.batch_norm:
self.batch_norm = self.batch_norm.cuda()
if self.cov_dropout:
self.cov_dropout = self.cov_dropout.cuda()
if self.activation:
self.activation = self.activation.cuda()
def forward(self, string, string_len):
""" process inputs
Args:
@ -102,7 +132,7 @@ class Conv(BaseLayer):
Tensor: shape: [batch_size, (seq_len - conv_window_size) // stride + 1, output_channel_num]
"""
if string_len is not None:
if string_len is not None and self.layer_conf.mask:
string_len_val = string_len.cpu().data.numpy()
masks = []
for i in range(len(string_len)):
@ -113,17 +143,21 @@ class Conv(BaseLayer):
masks = masks.to(device)
string = string * masks
string = torch.unsqueeze(string, 1) # [batch_size, input_channel_num=1, seq_len, feature_dim]
string_out = F.conv2d(string, self.filters[0], stride=self.layer_conf.stride, padding=self.layer_conf.padding)
if hasattr(self, 'batch_norms') and self.batch_norm:
string_out = self.batch_norm(string_out)
string_out = torch.squeeze(string_out, 3).permute(0, 2, 1)
string_ = string.transpose(2, 1).contiguous()
string_out = self.conv(string_)
if self.activation:
string_out = self.activation(string_out)
if string_len is not None:
string_len_out = (string_len - self.layer_conf.window_size) // self.layer_conf.stride + 1
else:
if self.cov_dropout:
string_out = self.cov_dropout(string_out)
if self.batch_norm:
string_out = self.batch_norm(string_out)
string_out = string_out.transpose(2, 1).contiguous()
string_len_out = None
if string_len is not None and self.layer_conf.remind_lengths:
string_len_out = string_len
return string_out, string_len_out

Просмотреть файл

@ -67,6 +67,9 @@ class EmbeddingConf(BaseConf):
for emb_type in self.conf:
if emb_type == 'position':
continue
if isinstance(self.conf[emb_type]['dim'], list):
self.output_dim[2] += sum(self.conf[emb_type]['dim'])
else:
self.output_dim[2] += self.conf[emb_type]['dim']
super(EmbeddingConf, self).inference()
@ -114,6 +117,7 @@ class Embedding(BaseLayer):
self.layer_conf = layer_conf
self.embeddings = nn.ModuleDict() if layer_conf.weight_on_gpu else dict()
self.char_embeddings = nn.ModuleDict()
for input_cluster in layer_conf.conf:
if 'type' in layer_conf.conf[input_cluster]:
# char embedding
@ -123,7 +127,7 @@ class Embedding(BaseLayer):
char_emb_conf = eval(layer_conf.conf[input_cluster]['type'] + "Conf")(** char_emb_conf_dict)
char_emb_conf.inference()
char_emb_conf.verify()
self.embeddings[input_cluster] = eval(layer_conf.conf[input_cluster]['type'])(char_emb_conf)
self.char_embeddings[input_cluster] = eval(layer_conf.conf[input_cluster]['type'])(char_emb_conf)
else:
# word embedding, postag embedding, and so on
self.embeddings[input_cluster] = nn.Embedding(layer_conf.conf[input_cluster]['vocab_size'], layer_conf.conf[input_cluster]['dim'], padding_idx=0)
@ -135,7 +139,6 @@ class Embedding(BaseLayer):
self.embeddings[input_cluster].weight.requires_grad = False
logging.info("The Embedding[%s][fix_weight] is true, fix the embeddings[%s]'s weight" % (input_cluster, input_cluster))
def forward(self, inputs, use_gpu=False):
""" process inputs
@ -157,11 +160,10 @@ class Embedding(BaseLayer):
if 'extra' in input_cluster:
continue
input = inputs[input_cluster]
# if 'type' in self.layer_conf.conf[input_cluster]:
# emb = self.embeddings[input_cluster](input, lengths[input]).float()
# else:
# emb = self.embeddings[input_cluster](input).float()
if self.embeddings[input_cluster].weight.device.type == 'cpu':
if input_cluster == 'char':
emb = self.char_embeddings[input_cluster](input).float()
else:
if list(self.embeddings[input_cluster].parameters())[0].device.type == 'cpu':
emb = self.embeddings[input_cluster](input.cpu()).float()
else:
emb = self.embeddings[input_cluster](input).float()

Просмотреть файл

@ -32,6 +32,7 @@ class LinearConf(BaseConf):
self.activation = 'PReLU'
self.last_hidden_activation = True
self.last_hidden_softmax = False
self.keep_dim = True # for exmaple if the output shape is [?, len, 1]. you want to squeeze it, set keep_dim=False, the the output shape is [?, len]
@DocInherit
def declare(self):
@ -42,9 +43,15 @@ class LinearConf(BaseConf):
def inference(self):
if isinstance(self.hidden_dim, int):
self.output_dim = copy.deepcopy(self.input_dims[0])
if not self.keep_dim and self.hidden_dim == 1:
self.output_dim.pop()
else:
self.output_dim[-1] = self.hidden_dim
elif isinstance(self.hidden_dim, list):
self.output_dim = copy.deepcopy(self.input_dims[0])
if not self.keep_dim and self.hidden_dim[-1] == 1:
self.output_dim.pop()
else:
self.output_dim[-1] = self.hidden_dim[-1]
super(LinearConf, self).inference() # PUT THIS LINE AT THE END OF inference()
@ -87,6 +94,7 @@ class Linear(BaseLayer):
def __init__(self, layer_conf):
super(Linear, self).__init__(layer_conf)
self.layer_conf = layer_conf
if layer_conf.input_ranks[0] == 3 and layer_conf.batch_norm is True:
layer_conf.batch_norm = False
@ -139,6 +147,8 @@ class Linear(BaseLayer):
masks = masks.to(device)
string = string * masks
string_out = self.linear(string.float())
if not self.layer_conf.keep_dim:
string_out = torch.squeeze(string_out, -1)
return string_out, string_len

104
block_zoo/Pooling1D.py Normal file
Просмотреть файл

@ -0,0 +1,104 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import torch
import torch.nn as nn
import torch.nn.functional as F
import numpy as np
from block_zoo.BaseLayer import BaseLayer, BaseConf
from utils.DocInherit import DocInherit
class Pooling1DConf(BaseConf):
"""
Args:
pool_type (str): 'max' or 'mean', default is 'max'.
stride (int): which axis to conduct pooling, default is 1.
padding (int): implicit zero paddings on both sides of the input. Can be a single number or a tuple (padH, padW). Default: 0
window_size (int): the size of the pooling
"""
def __init__(self, **kwargs):
super(Pooling1DConf, self).__init__(**kwargs)
@DocInherit
def default(self):
self.pool_type = 'max' # Supported: ['max', mean']
self.stride = 1
self.padding = 0
self.window_size = 3
@DocInherit
def declare(self):
self.num_of_inputs = 1
self.input_ranks = [3]
@DocInherit
def inference(self):
self.output_dim = [self.input_dims[0][0]]
if self.input_dims[0][1] != -1:
self.output_dim.append(
(self.input_dims[0][1] + 2 * self.padding - self.window_size) // self.stride + 1)
else:
self.output_dim.append(-1)
self.output_dim.append(self.input_dims[0][-1])
# DON'T MODIFY THIS
self.output_rank = len(self.output_dim)
@DocInherit
def verify(self):
super(Pooling1DConf, self).verify()
necessary_attrs_for_user = ['pool_type']
for attr in necessary_attrs_for_user:
self.add_attr_exist_assertion_for_user(attr)
self.add_attr_value_assertion('pool_type', ['max', 'mean'])
assert self.output_dim[
-1] != -1, "The shape of input is %s , and the input channel number of pooling should not be -1." % (
str(self.input_dims[0]))
class Pooling1D(BaseLayer):
""" Pooling layer
Args:
layer_conf (PoolingConf): configuration of a layer
"""
def __init__(self, layer_conf):
super(Pooling1D, self).__init__(layer_conf)
self.pool = None
if layer_conf.pool_type == "max":
self.pool = nn.MaxPool1d(kernel_size=layer_conf.window_size, stride=layer_conf.stride,
padding=layer_conf.padding)
elif layer_conf.pool_type == "mean":
self.pool = nn.AvgPool1d(kernel_size=layer_conf.window_size, stride=layer_conf.stride,
padding=layer_conf.padding)
def forward(self, string, string_len=None):
""" process inputs
Args:
string (Tensor): tensor with shape: [batch_size, length, feature_dim]
string_len (Tensor): [batch_size], default is None.
Returns:
Tensor: Pooling result of string
"""
string = string.permute([0, 2, 1]).contiguous()
string = self.pool(string)
string = string.permute([0, 2, 1]).contiguous()
return string, string_len

Просмотреть файл

@ -19,7 +19,6 @@ class Pooling2DConf(BaseConf):
stride (int): which axis to conduct pooling, default is 1.
padding (int): implicit zero paddings on both sides of the input. Can be a single number or a tuple (padH, padW). Default: 0
window_size (int): the size of the pooling
activation (string): activation functions, e.g. ReLU
"""
def __init__(self, **kwargs):
@ -30,7 +29,7 @@ class Pooling2DConf(BaseConf):
self.pool_type = 'max' # Supported: ['max', mean']
self.stride = 1
self.padding = 0
self.window_size = 3
# self.window_size = [self.input_dims[0][1], self.input_dims[0][2]]
@DocInherit
def declare(self):
@ -50,6 +49,9 @@ class Pooling2DConf(BaseConf):
@DocInherit
def inference(self):
if not hasattr(self, "window_size"):
self.window_size = [self.input_dims[0][1], self.input_dims[0][2]]
self.window_size = self.check_size(self.window_size, "window_size")
self.stride = self.check_size(self.stride, "stride")
self.padding = self.check_size(self.padding, "padding")

Просмотреть файл

@ -16,9 +16,13 @@ from .ConvPooling import ConvPooling, ConvPoolingConf
from .Dropout import Dropout, DropoutConf
from .Conv2D import Conv2D, Conv2DConf
from .Pooling1D import Pooling1D, Pooling1DConf
from .Pooling2D import Pooling2D, Pooling2DConf
from .embedding import CNNCharEmbedding, CNNCharEmbeddingConf
from .embedding import LSTMCharEmbedding, LSTMCharEmbeddingConf
from .CRF import CRFConf, CRF
from .attentions import FullAttention, FullAttentionConf
from .attentions import Seq2SeqAttention, Seq2SeqAttentionConf
@ -50,3 +54,4 @@ from .EncoderDecoder import EncoderDecoder, EncoderDecoderConf
from .normalizations import LayerNorm, LayerNormConf
from .HighwayLinear import HighwayLinear, HighwayLinearConf

Просмотреть файл

@ -58,7 +58,7 @@ class InteractionConf(BaseConf):
def verify(self):
super(InteractionConf, self).verify()
assert hasattr(self, 'matching_type'), "Please define matching_type attribute of BiGRUConf in default() or the configuration file"
assert self.matching_type in ['general', 'dot', 'mul', 'plus', 'minus', 'add'], "Invalid `matching_type`{self.matching_type} received. Must be in `mul`, `general`, `plus`, `minus`, `dot` and `concat`."
assert self.matching_type in ['general', 'dot', 'mul', 'plus', 'minus', 'add', 'concat'], "Invalid `matching_type`{self.matching_type} received. Must be in `mul`, `general`, `plus`, `minus`, `dot` and `concat`."
class Interaction(BaseLayer):
@ -120,7 +120,7 @@ class Interaction(BaseLayer):
return x - y
elif self.matching_type == 'concat':
def func(x, y):
return torch.concat([x, y], axis=-1)
return torch.cat([x, y], dim=-1)
else:
raise ValueError(f"Invalid matching type."
f"{self.matching_type} received."

Просмотреть файл

@ -28,11 +28,11 @@ class CNNCharEmbeddingConf(BaseConf):
@DocInherit
def default(self):
self.dim = 30 # cnn's output channel dim
self.dim = [30] # cnn's output channel dim
self.embedding_matrix_dim = 30 #
self.stride = 1
self.stride = [1]
self.padding = 0
self.window_size = 3
self.window_size = [3]
self.activation = 'ReLU'
@DocInherit
@ -41,8 +41,14 @@ class CNNCharEmbeddingConf(BaseConf):
self.num_of_inputs = 1
self.input_ranks = [3]
def change_to_list(self, attribute):
for single in attribute:
if not isinstance(getattr(self, single), list):
setattr(self, single, [getattr(self, single)])
@DocInherit
def inference(self):
self.change_to_list(['dim', 'stride', 'window_size'])
self.output_channel_num = self.dim
self.output_rank = 3
@ -65,20 +71,24 @@ class CNNCharEmbedding(BaseLayer):
super(CNNCharEmbedding, self).__init__(layer_conf)
self.layer_conf = layer_conf
assert len(layer_conf.dim) == len(layer_conf.window_size) == len(layer_conf.stride), "The attribute dim/window_size/stride must have the same length."
self.char_embeddings = nn.Embedding(layer_conf.vocab_size, layer_conf.embedding_matrix_dim, padding_idx=self.layer_conf.padding)
nn.init.uniform_(self.char_embeddings.weight, -0.001, 0.001)
self.filters = Variable(torch.randn(layer_conf.output_channel_num, layer_conf.input_channel_num,
layer_conf.window_size, layer_conf.embedding_matrix_dim).float(),
requires_grad=True)
self.char_cnn = nn.ModuleList()
for i in range(len(layer_conf.output_channel_num)):
self.char_cnn.append(nn.Conv2d(1, layer_conf.output_channel_num[i], (layer_conf.window_size[i], layer_conf.embedding_matrix_dim),
stride=self.layer_conf.stride[i], padding=self.layer_conf.padding))
if layer_conf.activation:
self.activation = eval("nn." + self.layer_conf.activation)()
else:
self.activation = None
if self.is_cuda():
self.filters = self.filters.cuda()
if self.activation:
self.activation.weight = torch.nn.Parameter(self.activation.weight.cuda())
# if self.is_cuda():
# self.char_embeddings = self.char_embeddings.cuda()
# self.char_cnn = self.char_cnn.cuda()
# if self.activation and hasattr(self.activation, 'weight'):
# self.activation.weight = torch.nn.Parameter(self.activation.weight.cuda())
def forward(self, string):
"""
@ -97,24 +107,29 @@ class CNNCharEmbedding(BaseLayer):
"""
string_reshaped = string.view(string.size()[0], -1) #[batch_size, seq_len * char num in words]
char_embs_lookup = self.char_embeddings(string_reshaped).float() # [batch_size, seq_len * char num in words, embedding_dim]
if self.is_cuda():
if self.filters.device == torch.device('cpu'):
self.filters = self.filters.cuda()
char_embs_lookup = char_embs_lookup.cuda(device=self.filters.device)
char_embs_lookup = char_embs_lookup.view(-1, string.size()[2], self.layer_conf.embedding_matrix_dim) #[batch_size * seq_len, char num in words, embedding_dim]
string_input = torch.unsqueeze(char_embs_lookup, 1) # [batch_size * seq_len, input_channel_num=1, char num in words, embedding_dim]
string_conv = F.conv2d(string_input, self.filters, stride=self.layer_conf.stride, padding=self.layer_conf.padding) # [batch_size * seq_len, output_channel_num, char num in word related, 1]
string_conv = torch.squeeze(string_conv, 3).permute(0, 2, 1) # [batch_size * seq_len, char num in word related, output_channel_num]
outputs = []
for index, single_cnn in enumerate(self.char_cnn):
string_conv = single_cnn(string_input).squeeze(3)
if self.activation:
string_conv = self.activation(string_conv)
string_maxpooling = torch.max(string_conv, 1)[0]
string_out = string_maxpooling.view(string.size()[0], string.size()[1], -1)
string_maxpooling = F.max_pool1d(string_conv, string_conv.size(2)).squeeze()
string_out = string_maxpooling.view(string.size()[0], -1, self.layer_conf.output_channel_num[index])
return string_out.cpu()
outputs.append(string_out)
if len(outputs) > 1:
string_output = torch.cat(outputs, 2)
else:
string_output = outputs[0]
return string_output
if __name__ == '__main__':

Просмотреть файл

@ -0,0 +1,137 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import torch
import torch.nn as nn
import torch.nn.functional as F
from torch.autograd import Variable
import numpy as np
from block_zoo.BaseLayer import BaseLayer, BaseConf
from utils.DocInherit import DocInherit
class LSTMCharEmbeddingConf(BaseConf):
""" Configuration of LSTMCharEmbedding
Args:
dim (int, optional): the dimension of character embedding after lstm. Default: 50
embedding_matrix_dim(int, optional): the dimension of character initialized embedding. Default: 30
padding(int, optional): Zero-padding added to both sides of the input. Default: 0
dropout(float, optional): dropout rate. Default: 0.2
bidirect_flag(Bool, optional): Using BiLSTM or not. Default: True
"""
def __init__(self, **kwargs):
super(LSTMCharEmbeddingConf, self).__init__(**kwargs)
@DocInherit
def default(self):
self.dim = 50 # lstm's output channel dim
self.embedding_matrix_dim = 30
self.padding = 0
self.dropout = 0.2
self.bidirect_flag = True
@DocInherit
def declare(self):
#self.input_channel_num = 1
self.num_of_inputs = 1
self.input_ranks = [3]
@DocInherit
def inference(self):
#self.output_channel_num = self.hidden_dim
self.output_rank = 3
@DocInherit
def verify(self):
# super(LSTMCharEmbeddingConf, self).verify()
necessary_attrs_for_user = ['embedding_matrix_dim', 'dim', 'dropout', 'bidirect_flag', 'vocab_size']
for attr in necessary_attrs_for_user:
self.add_attr_exist_assertion_for_user(attr)
class LSTMCharEmbedding(BaseLayer):
"""
This layer implements the character embedding use LSTM
Args:
layer_conf (LSTMCharEmbeddingConf): configuration of LSTMCharEmbedding
"""
def __init__(self, layer_conf):
super(LSTMCharEmbedding, self).__init__(layer_conf)
self.layer_conf = layer_conf
self.char_embeddings = nn.Embedding(layer_conf.vocab_size, layer_conf.embedding_matrix_dim, padding_idx=self.layer_conf.padding)
nn.init.uniform_(self.char_embeddings.weight, -0.001, 0.001)
if layer_conf.bidirect_flag:
self.dim = layer_conf.dim // 2
self.dropout = nn.Dropout(layer_conf.dropout)
self.char_lstm = nn.LSTM(layer_conf.embedding_matrix_dim, self.dim, num_layers=1, batch_first=True, bidirectional=layer_conf.bidirect_flag)
if self.is_cuda():
self.char_embeddings = self.char_embeddings.cuda()
self.dropout = self.dropout.cuda()
self.char_lstm = self.char_lstm.cuda()
def forward(self, string):
"""
Step1: [batch_size, seq_len, char num in words] -> [batch_size*seq_len, char num in words]
Step2: lookup embedding matrix -> [batch_size*seq_len, char num in words, embedding_dim]
Step3: after lstm operation, got [num_layer* num_directions, batch_size * seq_len, dim]
Step5: reshape -> [batch_size, seq_len, dim]
Args:
string (Variable): [[char ids of word1], [char ids of word2], [...], ...], shape: [batch_size, seq_len, char num in words]
Returns:
Variable: [batch_size, seq_len, output_dim]
"""
#print ('string shape: ', string.size())
string_reshaped = string.view(string.size()[0]*string.size()[1], -1) #[batch_size, seq_len * char num in words]
char_embs_lookup = self.char_embeddings(string_reshaped).float() # [batch_size, seq_len * char num in words, embedding_dim]
char_embs_drop = self.dropout(char_embs_lookup)
char_hidden = None
char_rnn_out, char_hidden = self.char_lstm(char_embs_drop, char_hidden)
#print('char_hidden shape: ', char_hidden[0].size())
string_out = char_hidden[0].transpose(1,0).contiguous().view(string.size()[0], string.size()[1], -1)
#print('string_out shape: ', string_out.size())
return string_out
if __name__ == '__main__':
conf = {
'embedding_matrix_dim': 30,
'dim': 30, # lstm's output channel dim
'padding': 0,
'dropout': 0.2,
'bidirect_flag': True,
# should be infered from the corpus
'vocab_size': 10,
'input_dims': [5],
'input_ranks': [3],
'use_gpu': True
}
layer_conf = LSTMCharEmbeddingConf(**conf)
# make a fake input: [bs, seq_len, char num in words]
# assume in this batch, the padded sentence length is 3 and the each word has 5 chars, including padding 0.
input_chars = np.array([
[[3, 1, 2, 5, 4], [1, 2, 3, 4, 0], [0, 0, 0, 0, 0]],
[[1, 1, 0, 0, 0], [2, 3, 1, 0, 0], [1, 2, 3, 4, 5]]
])
char_emb_layer = LSTMCharEmbedding(layer_conf)
input_chars = torch.LongTensor(input_chars)
output = char_emb_layer(input_chars)
print(output)

Просмотреть файл

@ -1,3 +1,4 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
from .CNNCharEmbedding import CNNCharEmbeddingConf, CNNCharEmbedding
from .LSTMCharEmbedding import LSTMCharEmbeddingConf, LSTMCharEmbedding

Просмотреть файл

@ -0,0 +1,97 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import torch
import torch.nn as nn
import torch.nn.functional as F
import logging
from ..BaseLayer import BaseConf, BaseLayer
from utils.DocInherit import DocInherit
from utils.exceptions import ConfigurationError
import copy
class CalculateDistanceConf(BaseConf):
""" Configuration of CalculateDistance Layer
Args:
operations (list): a subset of ["cos", "euclidean", "manhattan", "chebyshev"].
"""
# init the args
def __init__(self, **kwargs):
super(CalculateDistanceConf, self).__init__(**kwargs)
# set default params
@DocInherit
def default(self):
self.operations = ["cos", "euclidean", "manhattan", "chebyshev"]
@DocInherit
def declare(self):
self.num_of_inputs = 2
self.input_ranks = [2]
@DocInherit
def inference(self):
self.output_dim = copy.deepcopy(self.input_dims[0])
self.output_dim[-1] = 1
super(CalculateDistanceConf, self).inference()
@DocInherit
def verify(self):
super(CalculateDistanceConf, self).verify()
assert len(self.input_dims) == 2, "Operation requires that there should be two inputs"
# to check if the ranks of all the inputs are equal
rank_equal_flag = True
for i in range(len(self.input_ranks)):
if self.input_ranks[i] != self.input_ranks[0] or self.input_ranks[i] != 2:
rank_equal_flag = False
break
if rank_equal_flag == False:
raise ConfigurationError("For layer CalculateDistance, the ranks of each inputs should be equal and 2!")
class CalculateDistance(BaseLayer):
""" CalculateDistance layer to calculate the distance of sequences(2D representation)
Args:
layer_conf (CalculateDistanceConf): configuration of a layer
"""
def __init__(self, layer_conf):
super(CalculateDistance, self).__init__(layer_conf)
self.layer_conf = layer_conf
def forward(self, x, x_len, y, y_len):
"""
Args:
x: [batch_size, dim]
x_len: [batch_size]
y: [batch_size, dim]
y_len: [batch_size]
Returns:
Tensor: [batch_size, 1], None
"""
batch_size = x.size()[0]
if "cos" in self.layer_conf.operations:
result = F.cosine_similarity(x , y)
elif "euclidean" in self.layer_conf.operations:
result = torch.sqrt(torch.sum((x-y)**2, dim=1))
elif "manhattan" in self.layer_conf.operations:
result = torch.sum(torch.abs((x - y)), dim=1)
elif "chebyshev" in self.layer_conf.operations:
result = torch.abs((x - y)).max(dim=1)
else:
raise ConfigurationError("This operation is not supported!")
result = result.view(batch_size, 1)
return result, None

Просмотреть файл

@ -47,7 +47,6 @@ class CombinationConf(BaseConf):
self.output_dim[-1] += int(np.mean([input_dim[-1] for input_dim in self.input_dims])) # difference operation requires dimension of all the inputs should be equal
if "dot_multiply" in self.operations:
self.output_dim[-1] += int(np.mean([input_dim[-1] for input_dim in self.input_dims])) # dot_multiply operation requires dimension of all the inputs should be equal
super(CombinationConf, self).inference()
@DocInherit

Просмотреть файл

@ -0,0 +1,76 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
# Come from http://www.hangli-hl.com/uploads/3/1/6/8/3168008/hu-etal-nips2014.pdf [ARC-II]
import torch
import torch.nn as nn
import copy
from block_zoo.BaseLayer import BaseLayer, BaseConf
from utils.DocInherit import DocInherit
from utils.exceptions import ConfigurationError
class Expand_plusConf(BaseConf):
"""Configuration for Expand_plus layer
"""
def __init__(self, **kwargs):
super(Expand_plusConf, self).__init__(**kwargs)
@DocInherit
def default(self):
self.operation = 'Plus'
@DocInherit
def declare(self):
self.num_of_inputs = 2
self.input_ranks = [3, 3]
@DocInherit
def inference(self):
self.output_dim = copy.deepcopy(self.input_dims[0])
if self.input_dims[0][1] == -1 or self.input_dims[1][1] == -1:
raise ConfigurationError("For Expand_plus layer, the sequence length should be fixed")
self.output_dim.insert(2, self.input_dims[1][1]) # y_len
super(Expand_plusConf, self).inference() # PUT THIS LINE AT THE END OF inference()
@DocInherit
def verify(self):
super(Expand_plusConf, self).verify()
class Expand_plus(BaseLayer):
""" Expand_plus layer
Given sequences X and Y, put X and Y expand_dim, and then add.
Args:
layer_conf (Expand_plusConf): configuration of a layer
"""
def __init__(self, layer_conf):
super(Expand_plus, self).__init__(layer_conf)
assert layer_conf.input_dims[0][-1] == layer_conf.input_dims[1][-1]
def forward(self, x, x_len, y, y_len):
"""
Args:
x: [batch_size, x_max_len, dim].
x_len: [batch_size], default is None.
y: [batch_size, y_max_len, dim].
y_len: [batch_size], default is None.
Returns:
output: batch_size, x_max_len, y_max_len, dim].
"""
x_new = torch.stack([x]*y.size()[1], 2) # [batch_size, x_max_len, y_max_len, dim]
y_new = torch.stack([y]*x.size()[1], 1) # [batch_size, x_max_len, y_max_len, dim]
return x_new + y_new, None

Просмотреть файл

@ -5,3 +5,5 @@ from .Concat3D import Concat3D, Concat3DConf
from .Combination import Combination, CombinationConf
from .Match import Match, MatchConf
from .Flatten import Flatten, FlattenConf
from .Expand_plus import Expand_plus, Expand_plusConf
from .CalculateDistance import CalculateDistance, CalculateDistanceConf

8
dataset/get_glove.sh Normal file → Executable file
Просмотреть файл

@ -1,13 +1,21 @@
preprocess_exec="sed -f tokenizer.sed"
glovepath='http://nlp.stanford.edu/data/glove.840B.300d.zip'
glovepath_6B='http://nlp.stanford.edu/data/glove.6B.zip'
ZIPTOOL="unzip"
# GloVe
echo $glovepath
if [ ! -d "/GloVe/"];then
mkdir GloVe
fi
curl -LO $glovepath
$ZIPTOOL glove.840B.300d.zip -d GloVe/
rm glove.840B.300d.zip
curl -LO $glovepath_6B
$ZIPTOOL glove.6B.zip -d GloVe/
rm glove.6B.zip

15
dataset/get_glove_6B.sh Normal file
Просмотреть файл

@ -0,0 +1,15 @@
preprocess_exec="sed -f tokenizer.sed"
glovepath='http://nlp.stanford.edu/data/glove.6B.zip'
ZIPTOOL="unzip"
# GloVe
echo $glovepath
if [ ! -d "/GloVe/"];then
mkdir GloVe
fi
curl -LO $glovepath
$ZIPTOOL glove.6B.zip -d GloVe/
rm glove.6B.zip

Просмотреть файл

@ -0,0 +1,500 @@
-DOCSTART- O
CRICKET - LEICESTERSHIRE TAKE OVER AT TOP AFTER INNINGS VICTORY . O O S-ORG O O O O O O O O
LONDON 1996-08-30 S-LOC O
West Indian all-rounder Phil Simmons took four for 38 on Friday as Leicestershire beat Somerset by an innings and 39 runs in two days to take over at the head of the county championship . B-MISC E-MISC O B-PER E-PER O O O O O O O S-ORG O S-ORG O O O O O O O O O O O O O O O O O O O O
Their stay on top , though , may be short-lived as title rivals Essex , Derbyshire and Surrey all closed in on victory while Kent made up for lost time in their rain-affected match against Nottinghamshire . O O O O O O O O O O O O O S-ORG O S-ORG O S-ORG O O O O O O S-ORG O O O O O O O O O O S-ORG O
After bowling Somerset out for 83 on the opening morning at Grace Road , Leicestershire extended their first innings by 94 runs before being bowled out for 296 with England discard Andy Caddick taking three for 83 . O O S-ORG O O O O O O O O B-LOC E-LOC O S-ORG O O O O O O O O O O O O O O S-LOC O B-PER E-PER O O O O O
Trailing by 213 , Somerset got a solid start to their second innings before Simmons stepped in to bundle them out for 174 . O O O O S-ORG O O O O O O O O O S-PER O O O O O O O O O
Essex , however , look certain to regain their top spot after Nasser Hussain and Peter Such gave them a firm grip on their match against Yorkshire at Headingley . S-ORG O O O O O O O O O O O B-PER E-PER O B-PER E-PER O O O O O O O O O S-ORG O S-LOC O
Hussain , considered surplus to England 's one-day requirements , struck 158 , his first championship century of the season , as Essex reached 372 and took a first innings lead of 82 . S-PER O O O O S-LOC O O O O O O O O O O O O O O O O S-ORG O O O O O O O O O O O
By the close Yorkshire had turned that into a 37-run advantage but off-spinner Such had scuttled their hopes , taking four for 24 in 48 balls and leaving them hanging on 119 for five and praying for rain . O O O S-ORG O O O O O O O O O S-PER O O O O O O O O O O O O O O O O O O O O O O O O O
At the Oval , Surrey captain Chris Lewis , another man dumped by England , continued to silence his critics as he followed his four for 45 on Thursday with 80 not out on Friday in the match against Warwickshire . O O S-LOC O S-ORG O B-PER E-PER O O O O O S-LOC O O O O O O O O O O O O O O O O O O O O O O O O O S-ORG O
He was well backed by England hopeful Mark Butcher who made 70 as Surrey closed on 429 for seven , a lead of 234 . O O O O O S-LOC O B-PER E-PER O O O O S-ORG O O O O O O O O O O O
Derbyshire kept up the hunt for their first championship title since 1936 by reducing Worcestershire to 133 for five in their second innings , still 100 runs away from avoiding an innings defeat . S-ORG O O O O O O O O O O O O O S-ORG O O O O O O O O O O O O O O O O O O O
Australian Tom Moody took six for 82 but Chris Adams , 123 , and Tim O'Gorman , 109 , took Derbyshire to 471 and a first innings lead of 233 . S-MISC B-PER E-PER O O O O O B-PER E-PER O O O O B-PER E-PER O O O O S-ORG O O O O O O O O O O
After the frustration of seeing the opening day of their match badly affected by the weather , Kent stepped up a gear to dismiss Nottinghamshire for 214 . O O O O O O O O O O O O O O O O O S-ORG O O O O O O S-ORG O O O
They were held up by a gritty 84 from Paul Johnson but ex-England fast bowler Martin McCague took four for 55 . O O O O O O O O O B-PER E-PER O S-MISC O O B-PER E-PER O O O O O
By stumps Kent had reached 108 for three . O O S-ORG O O O O O O
-DOCSTART- O
CRICKET - ENGLISH COUNTY CHAMPIONSHIP SCORES . O O B-MISC I-MISC E-MISC O O
LONDON 1996-08-30 S-LOC O
Result and close of play scores in English county championship matches on Friday : O O O O O O O S-MISC O O O O O O
Leicester : Leicestershire beat Somerset by an innings and 39 runs . S-LOC O S-ORG O S-ORG O O O O O O O
Somerset 83 and 174 ( P. Simmons 4-38 ) , Leicestershire 296 . S-ORG O O O O B-PER E-PER O O O S-ORG O O
Leicestershire 22 points , Somerset 4 . S-ORG O O O S-ORG O O
Chester-le-Street : Glamorgan 259 and 207 ( A. Dale 69 , H. Morris 69 ; D. Blenkiron 4-43 ) , Durham 114 ( S. Watkin 4-28 ) and 81-3 . S-LOC O S-ORG O O O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O O O S-ORG O O B-PER E-PER O O O O O
Tunbridge Wells : Nottinghamshire 214 ( P. Johnson 84 ; M. McCague 4-55 ) , Kent 108-3 . B-LOC E-LOC O S-ORG O O B-PER E-PER O O B-PER E-PER O O O S-ORG O O
London ( The Oval ) : Warwickshire 195 , Surrey 429-7 ( C. Lewis 80 not out , M. Butcher 70 , G. Kersey 63 , J. Ratcliffe 63 , D. Bicknell 55 ) . S-LOC O B-LOC E-LOC O O S-ORG O O S-ORG O O B-PER E-PER O O O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O O O
Hove : Sussex 363 ( W. Athey 111 , V. Drakes 52 ; I. Austin 4-37 ) , Lancashire 197-8 ( W. Hegg 54 ) S-LOC O S-ORG O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O O O S-ORG O O B-PER E-PER O O
Portsmouth : Middlesex 199 and 426 ( J. Pooley 111 , M. Ramprakash 108 , M. Gatting 83 ) , Hampshire 232 and 109-5 . S-LOC O S-ORG O O O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O O O S-ORG O O O O
Chesterfield : Worcestershire 238 and 133-5 , Derbyshire 471 ( J. Adams 123 , T.O'Gorman 109 not out , K. Barnett 87 ; T. Moody 6-82 ) S-LOC O S-ORG O O O O S-ORG O O B-PER E-PER O O S-PER O O O O B-PER E-PER O O B-PER E-PER O O
Bristol : Gloucestershire 183 and 185-6 ( J. Russell 56 not out ) , Northamptonshire 190 ( K. Curran 52 ; A. Smith 5-68 ) . S-LOC O S-ORG O O O O B-PER E-PER O O O O O S-ORG O O B-PER E-PER O O B-PER E-PER O O O
-DOCSTART- O
CRICKET - 1997 ASHES INTINERARY . O O O S-MISC O O
LONDON 1996-08-30 S-LOC O
Australia will defend the Ashes in S-LOC O O O S-MISC O
a six-test series against England during a four-month tour O O O O S-LOC O O O O
starting on May 13 next year , the Test and County Cricket Board O O O O O O O O B-ORG I-ORG I-ORG I-ORG E-ORG
said on Friday . O O O O
Australia will also play three one-day internationals and S-LOC O O O O O O O
four one-day warm-up matches at the start of the tour . O O O O O O O O O O O
The tourists will play nine first-class matches against O O O O O O O O
English county sides and another against British Universities , S-MISC O O O O O B-ORG E-ORG O
as well as one-day matches against the Minor Counties and O O O O O O O B-ORG E-ORG O
Scotland . S-LOC O
Tour itinerary : O O O
May O
May 13 Arrive in London O O O O S-LOC
May 14 Practice at Lord 's O O O O B-LOC E-LOC
May 15 v Duke of Norfolk 's XI ( at Arundel ) O O O B-ORG I-ORG I-ORG I-ORG E-ORG O O S-LOC O
May 17 v Northampton O O O S-ORG
May 18 v Worcestershire O O O S-ORG
May 20 v Durham O O O S-ORG
May 22 First one-day international ( at Headingley , O O O O O O O S-LOC O
Leeds ) S-ORG O
May 24 Second one-day international ( at The Oval , O O O O O O O B-LOC E-LOC O
London ) S-LOC O
May 25 Third one-day international ( at Lord 's , London ) O O O O O O O B-LOC E-LOC O S-LOC O
May 27-29 v Gloucestershire or Sussex or Surrey ( three O O O S-ORG O S-ORG O S-ORG O O
days ) O O
May 31 - June 2 v Derbyshire ( three days ) O O O O O O S-ORG O O O O
June O
June 5-9 First test match ( at Edgbaston , Birmingham ) O O O O O O O S-LOC O S-LOC O
June 11-13 v a first class county ( to be confirmed ) O O O O O O O O O O O O
June 14-16 v Leicestershire ( three days ) O O O S-ORG O O O O
June 19-23 Second test ( at Lord 's ) O O O O O O B-LOC E-LOC O
June 25-27 v British Universities ( at Oxford , three days ) O O O B-ORG E-ORG O O S-LOC O O O O
June 28-30 v Hampshire ( three days ) O O O S-ORG O O O O
July O
July 3-7 Third test ( at Old Trafford , Manchester ) O O O O O O B-LOC E-LOC O S-LOC O
July 9 v Minor Counties XI O O O B-ORG I-ORG E-ORG
July 12 v Scotland O O O S-LOC
July 16-18 v Glamorgan ( three days ) O O O S-ORG O O O O
July 19-21 v Middlesex ( three days ) O O O S-ORG O O O O
July 24-28 Fourth test ( at Headingley ) O O O O O O S-LOC O
August O
August 1-4 v Somerset ( four days ) O O O S-ORG O O O O
August 7-11 Fifth test ( at Trent Bridge , Nottingham ) O O O O O O B-LOC E-LOC O S-LOC O
August 16-18 v Kent ( three days ) O O O S-ORG O O O O
August 21-25 Sixth test ( at The Oval , London ) . O O O O O O B-LOC E-LOC O S-LOC O O
-DOCSTART- O
SOCCER - SHEARER NAMED AS ENGLAND CAPTAIN . O O S-PER O O S-LOC O O
LONDON 1996-08-30 S-LOC O
The world 's costliest footballer Alan Shearer was named as the new England captain on Friday . O O O O O B-PER E-PER O O O O O S-LOC O O O O
The 26-year-old , who joined Newcastle for 15 million pounds sterling ( $ 23.4 million ) , takes over from Tony Adams , who led the side during the European championship in June , and former captain David Platt . O O O O O S-ORG O O O O O O O O O O O O O O B-PER E-PER O O O O O O O S-MISC O O O O O O O B-PER E-PER O
Adams and Platt are both injured and will miss England 's opening World Cup qualifier against Moldova on Sunday . S-PER O S-PER O O O O O O S-LOC O O B-MISC E-MISC O O S-LOC O O O
Shearer takes the captaincy on a trial basis , but new coach Glenn Hoddle said he saw no reason why the former Blackburn and Southampton skipper should not make the post his own . S-PER O O O O O O O O O O O B-PER E-PER O O O O O O O O S-ORG O S-ORG O O O O O O O O O
" I 'm sure there wo n't be a problem , I 'm sure Alan is the man for the job , " Hoddle said . O O O O O O O O O O O O O O S-PER O O O O O O O O S-PER O O
" There were three or four people who could have done it but when I spoke to Alan he was up for it and really wanted it . O O O O O O O O O O O O O O O O O S-PER O O O O O O O O O O
" In four days it 's very difficult to come to a 100 percent conclusion about something like this ... O O O O O O O O O O O O O O O O O O O O
but he knows how to conduct himself , his team mates respect him and he knows about the team situation even though he plays up front . " O O O O O O O O O O O O O O O O O O O O O O O O O O O O
Shearer 's Euro 96 striking partner Teddy Sheringham withdrew from the squad with an injury on Friday . S-PER O B-MISC E-MISC O O B-PER E-PER O O O O O O O O O O
He will probably be replaced by Shearer 's Newcastle team mate Les Ferdinand . O O O O O O S-PER O S-ORG O O B-PER E-PER O
-DOCSTART- O
BASKETBALL - INTERNATIONAL TOURNAMENT RESULT . O O O O O O
BELGRADE 1996-08-30 S-LOC O
Result in an international O O O O
basketball tournament on Friday : O O O O O
Red Star ( Yugoslavia ) beat Dinamo ( Russia ) 92-90 ( halftime B-ORG E-ORG O S-LOC O O S-ORG O S-LOC O O O O
47-47 ) O O
-DOCSTART- O
SOCCER - ROMANIA BEAT LITHUANIA IN UNDER-21 MATCH . O O S-LOC O S-LOC O O O O
BUCHAREST 1996-08-30 S-LOC O
Romania beat Lithuania 2-1 ( halftime 1-1 ) in their European under-21 soccer match on Friday . S-LOC O S-LOC O O O O O O O S-MISC O O O O O O
Scorers : O O
Romania - Cosmin Contra ( 31st ) , Mihai Tararache ( 75th ) S-LOC O B-PER E-PER O O O O B-PER E-PER O O O
Lithuania - Danius Gleveckas ( 13rd ) S-LOC O B-PER E-PER O O O
Attendance : 200 O O O
-DOCSTART- O
SOCCER - ROTOR FANS LOCKED OUT AFTER VOLGOGRAD VIOLENCE . O O S-ORG O O O O S-LOC O O
MOSCOW 1996-08-30 S-LOC O
Rotor Volgograd must play their next home game behind closed doors after fans hurled bottles and stones at Dynamo Moscow players during a 1-0 home defeat on Saturday that ended Rotor 's brief spell as league leaders . B-ORG E-ORG O O O O O O O O O O O O O O O O B-ORG E-ORG O O O O O O O O O O S-ORG O O O O O O O
The head of the Russian league 's disciplinary committee , Anatoly Gorokhovsky , said on Friday that Rotor would play Lada Togliatti to empty stands on September 3 . O O O O S-MISC O O O O O B-PER E-PER O O O O O S-ORG O O B-ORG E-ORG O O O O O O O
The club , who put Manchester United out of last year 's UEFA Cup , were fined $ 1,000 . O O O O O B-ORG E-ORG O O O O O B-MISC E-MISC O O O O O O
Despite the defeat , Rotor are well placed with 11 games to play in the championship . O O O O S-ORG O O O O O O O O O O O O
Lying three points behind Alania and two behind Dynamo Moscow , the Volgograd side have a game in hand over the leaders and two over the Moscow club . O O O O S-ORG O O O B-ORG E-ORG O O S-LOC O O O O O O O O O O O O O S-LOC O O
-DOCSTART- O
BOXING - PANAMA 'S ROBERTO DURAN FIGHTS THE SANDS OF TIME . O O S-LOC O B-PER E-PER O O O O O O
PANAMA CITY 1996-08-30 B-LOC E-LOC O
Panamanian boxing legend Roberto " Hands of Stone " Duran climbs into the ring on Saturday in another age-defying attempt to sustain his long career . S-MISC O O B-PER I-PER I-PER I-PER I-PER I-PER E-PER O O O O O O O O O O O O O O O O
Duran , 45 , takes on little-known Mexican Ariel Cruz , 30 , in a super middleweight non-title bout in Panama City . S-PER O O O O O O S-MISC B-PER E-PER O O O O O O O O O O B-LOC E-LOC O
The fight , Duran 's first on home soil for 10 years , is being billed here as the " Return of the Legend " and Duran still talks as if he was in his prime . O O O S-PER O O O O O O O O O O O O O O O O B-MISC I-MISC I-MISC E-MISC O O S-PER O O O O O O O O O O
" I want a fifth title . O O O O O O O
This match is to prepare me . O O O O O O O
I feel good . O O O O
I 'm not retiring , " Duran told Reuters . O O O O O O S-PER O S-ORG O
But those close to the boxer acknowledge that the man who has won championships in four different weight classes -- lightweight , welterweight , junior middleweight and middleweight -- is drawing close to the end of his career . O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O
" Each time he fights , he 's on the last frontier of his career . O O O O O O O O O O O O O O O O
If he loses Saturday , it could devalue his position as one of the world 's great boxers , " Panamanian Boxing Association President Ramon Manzanares said . O O O O O O O O O O O O O O O O O O O O S-MISC B-ORG E-ORG O B-PER E-PER O O
Duran , whose 97-12 record spans three decades , hopes a win in the 10-round bout will earn him a rematch against Puerto Rico 's Hector " Macho " Camacho . S-PER O O O O O O O O O O O O O O O O O O O O O B-LOC E-LOC O B-PER I-PER I-PER I-PER E-PER O
Camacho took a controversial points decision against the Panamanian in Atlantic City in June in a title fight . S-PER O O O O O O O S-MISC O B-LOC E-LOC O O O O O O O
-DOCSTART- O
SQUASH - HONG KONG OPEN QUARTER-FINAL RESULTS . O O B-MISC I-MISC E-MISC O O O
HONG KONG 1996-08-30 B-LOC E-LOC O
Quarter-final results in the Hong Kong Open on Friday ( prefix number denotes seeding ) : 1 - Jansher Khan ( Pakistan ) beat Mark Cairns ( England ) 15-10 15-6 15-7 O O O O B-MISC I-MISC E-MISC O O O O O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O
Anthony Hill ( Australia ) beat Dan Jenson ( Australia ) 15-9 15-8 15-17 17-15 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O
4 - Peter Nicol ( Scotland ) beat 7 - Chris Walker ( England ) 15-8 15-13 13-15 15-9 O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O O
2 - Rodney Eyles ( Australia ) beat Derek Ryan ( Ireland ) 15-6 15-9 11-15 15-10 . O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O
-DOCSTART- O
SOCCER - RESULTS OF SOUTH KOREAN PRO-SOCCER GAMES . O O O O B-MISC E-MISC O O O
SEOUL 1996-08-30 S-LOC O
Results of South Korean pro-soccer O O B-MISC E-MISC O
games played on Thursday . O O O O O
Pohang 3 Ulsan 2 ( halftime 1-0 ) S-ORG O S-ORG O O O O O
Puchon 2 Chonbuk 1 ( halftime 1-1 ) S-ORG O S-ORG O O O O O
Standings after games played on Thursday ( tabulate under - O O O O O O O O O O
won , drawn , lost , goals for , goals against , points ) : O O O O O O O O O O O O O O O
W D L G / F G / A P O O O O O O O O O O
Puchon 3 1 0 6 1 10 S-ORG O O O O O O
Chonan 3 0 1 13 10 9 S-ORG O O O O O O
Pohang 2 1 1 11 10 7 S-ORG O O O O O O
Suwan 1 3 0 7 3 6 S-ORG O O O O O O
Ulsan 1 0 2 8 9 3 S-ORG O O O O O O
Anyang 0 3 1 6 9 3 S-ORG O O O O O O
Chonnam 0 2 1 4 5 2 S-ORG O O O O O O
Pusan 0 2 1 3 7 2 S-ORG O O O O O O
Chonbuk 0 0 3 3 7 0 S-ORG O O O O O O
-DOCSTART- O
BASEBALL - RESULTS OF S. KOREAN PROFESSIONAL GAMES . O O O O B-MISC E-MISC O O O
SEOUL 1996-08-30 S-LOC O
Results of South Korean O O B-MISC E-MISC
professional baseball games played on Thursday . O O O O O O O
LG 2 OB 0 S-ORG O S-ORG O
Lotte 6 Hyundai 2 S-ORG O S-ORG O
Hyundai 6 Lotte 5 S-ORG O S-ORG O
Haitai 2 Samsung 0 S-ORG O S-ORG O
Samsung 10 Haitai 3 S-ORG O S-ORG O
Hanwha 6 Ssangbangwool 5 S-ORG O S-ORG O
Note - Lotte and Hyundai , Haitai and Samsung played two games . O O S-ORG O S-ORG O S-ORG O S-ORG O O O O
Standings after games played on Thursday ( tabulate under O O O O O O O O O
won , drawn , lost , winning percentage , games behind first place ) O O O O O O O O O O O O O O
W D L PCT GB O O O O O
Haitai 64 2 43 .596 - S-ORG O O O O O
Ssangbangwool 59 2 49 .545 5 1/2 S-ORG O O O O O O
Hanwha 58 1 49 .542 6 S-ORG O O O O O
Hyundai 57 5 49 .536 6 1/2 S-ORG O O O O O O
Samsung 49 5 56 .468 14 S-ORG O O O O O
Lotte 46 6 54 .462 14 1/2 S-ORG O O O O O O
LG 46 5 59 .441 17 S-ORG O O O O O
OB 42 6 62 .409 20 1/2 S-ORG O O O O O O
-DOCSTART- O
TENNIS - FRIDAY 'S RESULTS FROM THE U.S. OPEN . O O O O O O O B-MISC E-MISC O
NEW YORK 1996-08-30 B-LOC E-LOC O
Results from the U.S. Open Tennis Championships at the National Tennis Centre on Friday ( prefix number denotes seeding ) : O O O B-MISC I-MISC I-MISC E-MISC O O B-LOC I-LOC E-LOC O O O O O O O O O
Women 's singles , third round O O O O O O
Sandrine Testud ( France ) beat Ines Gorrochategui ( Argentina ) 4-6 6-2 6-1 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O
Men 's singles , second round O O O O O O
4 - Goran Ivanisevic ( Croatia ) beat Scott Draper ( Australia ) 6-7 ( 1-7 ) 6-3 6-4 6-4 O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O O O
Tim Henman ( Britain ) beat Doug Flach ( U.S. ) 6-3 6-4 6-2 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O
Mark Philippoussis ( Australia ) beat Andrei Olhovskiy ( Russia ) 6 - 3 6-4 6-2 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O
Sjeng Schalken ( Netherlands ) beat David Rikl ( Czech Republic ) 6 - 2 6-4 6-4 B-PER E-PER O S-LOC O O B-PER E-PER O B-LOC E-LOC O O O O O O
Guy Forget ( France ) beat 17 - Felix Mantilla ( Spain ) 6-4 7-5 6-3 B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O
Men 's singles , second round O O O O O O
Alexander Volkov ( Russia ) beat Mikael Tillstrom ( Sweden ) 1-6 6- 4 6-1 4-6 7-6 ( 10-8 ) B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O O O O O
Jonas Bjorkman ( Sweden ) beat David Nainkin ( South Africa ) ) 6-4 6-1 6-1 B-PER E-PER O S-LOC O O B-PER E-PER O B-LOC E-LOC O O O O O
Women 's singles , third round O O O O O O
8 - Lindsay Davenport ( U.S. ) beat Anne-Gaelle Sidot ( France ) 6-0 6-3 O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
4 - Conchita Martinez ( Spain ) beat Helena Sukova ( Czech Republic ) 6-4 6-3 O O B-PER E-PER O S-LOC O O B-PER E-PER O B-LOC E-LOC O O O
Amanda Coetzer ( South Africa ) beat Irina Spirlea ( Romania ) 7-6 ( 7-5 ) 7-5 B-PER E-PER O B-LOC E-LOC O O B-PER E-PER O S-LOC O O O O O O
Add Men 's singles , second round 16 - Cedric Pioline ( France ) beat Roberto Carretero ( Spain ) 4-6 6 - 2 6-2 6-1 Alex Corretja ( Spain ) beat Filippo Veglio ( Switzerland ) 6-7 ( 4- 7 ) 6-4 6-4 6-0 O O O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O O O O
Add Women 's singles , third round Linda Wild ( U.S. ) beat Barbara Rittner ( Germany ) 6-4 4-6 7-5 Asa Carlsson ( Sweden ) beat 15 - Gabriela Sabatini ( Argentina ) 7-5 3-6 6-2 O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O
Add Men 's singles , second round 1 - Pete Sampras ( U.S. ) beat Jiri Novak ( Czech Republic ) 6-3 1-6 6-3 4-6 6-4 Paul Haarhuis ( Netherlands ) beat Michael Tebbutt ( Australia ) 1- 6 6-2 6-2 6-3 O O O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O B-LOC E-LOC O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O
Add Women 's singles , third round Lisa Raymond ( U.S. ) beat Kimberly Po ( U.S. ) 6-3 6-2 O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
: O
Add men 's singles , second round O O O O O O O
Hendrik Dreekmann ( Germany ) beat Thomas Johansson ( Sweden ) B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O
7-6 ( 7-1 ) 6-2 4-6 6-1 O O O O O O O
Andrei Medvedev ( Ukraine ) beat Jan Kroslak ( Slovakia ) 6-4 6-3 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
6-2 O
Petr Korda ( Czech Republic ) bat Bohdan Ulihrach ( Czech B-PER E-PER O B-LOC E-LOC O O B-PER E-PER O S-LOC
Republic ) 6-0 7-6 ( 7-5 ) 6-2 S-LOC O O O O O O O
Add women 's singles , third round O O O O O O O
2 - Monica Seles ( U.S. ) beat Dally Randriantefy ( Madagascar ) O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O
6-0 6-2 O O
: O
Add men 's singles , second round 12 - Todd Martin ( U.S. ) beat Andrea Gaudenzi ( Italy ) 6-3 6-2 6-2 Stefan Edberg ( Sweden ) beat Bernd Karbacher ( Germany ) 3-6 6-3 6-3 1-0 retired ( leg injury ) O O O O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O O O O O O O
-DOCSTART- O
BASEBALL - MAJOR LEAGUE STANDINGS AFTER THURSDAY 'S GAMES . O O B-MISC E-MISC O O O O O O
NEW YORK 1996-08-30 B-LOC E-LOC O
Major League Baseball B-MISC I-MISC E-MISC
standings after games played on Thursday ( tabulate under won , O O O O O O O O O O O
lost , winning percentage and games behind ) : O O O O O O O O O
AMERICAN LEAGUE B-MISC E-MISC
EASTERN DIVISION B-MISC E-MISC
W L PCT GB O O O O
NEW YORK 74 59 .556 - B-ORG E-ORG O O O O
BALTIMORE 70 63 .526 4 S-ORG O O O O
BOSTON 69 65 .515 5 1/2 S-ORG O O O O O
TORONTO 63 71 .470 11 1/2 S-ORG O O O O O
DETROIT 48 86 .358 26 1/2 S-ORG O O O O O
CENTRAL DIVISION B-MISC E-MISC
CLEVELAND 80 53 .602 - S-ORG O O O O
CHICAGO 71 64 .526 10 S-ORG O O O O
MINNESOTA 67 67 .500 13 1/2 S-ORG O O O O O
MILWAUKEE 64 71 .474 17 S-ORG O O O O
KANSAS CITY 61 74 .452 20 B-ORG E-ORG O O O O
WESTERN DIVISION B-MISC E-MISC
TEXAS 75 58 .564 - S-ORG O O O O
SEATTLE 70 63 .526 5 S-ORG O O O O
OAKLAND 64 72 .471 12 1/2 S-ORG O O O O O
CALIFORNIA 62 72 .463 13 1/2 S-ORG O O O O O
FRIDAY , AUGUST 30 SCHEDULE O O O O O
KANSAS CITY AT DETROIT B-ORG E-ORG O S-LOC
CHICAGO AT TORONTO S-ORG O S-LOC
MINNESOTA AT MILWAUKEE S-ORG O S-LOC
CLEVELAND AT TEXAS S-ORG O S-LOC
NEW YORK AT CALIFORNIA B-ORG E-ORG O S-LOC
BOSTON AT OAKLAND S-ORG O S-LOC
BALTIMORE AT SEATTLE S-ORG O S-LOC
NATIONAL LEAGUE B-MISC E-MISC
EASTERN DIVISION B-MISC E-MISC
W L PCT GB O O O O
ATLANTA 83 49 .629 - S-ORG O O O O
MONTREAL 71 61 .538 12 S-ORG O O O O
FLORIDA 64 70 .478 20 S-ORG O O O O
NEW YORK 59 75 .440 25 B-ORG E-ORG O O O O
PHILADELPHIA 54 80 .403 30 S-ORG O O O O
CENTRAL DIVISION B-MISC E-MISC
HOUSTON 72 63 .533 - S-ORG O O O O
ST LOUIS 69 65 .515 2 1/2 B-ORG E-ORG O O O O O
CINCINNATI 66 67 .496 5 S-ORG O O O O
CHICAGO 65 66 .496 5 S-ORG O O O O
PITTSBURGH 56 77 .421 15 S-ORG O O O O
WESTERN DIVISION B-MISC E-MISC
SAN DIEGO 75 60 .556 - B-ORG E-ORG O O O O
LOS ANGELES 72 61 .541 2 B-ORG E-ORG O O O O
COLORADO 70 65 .519 5 S-ORG O O O O
SAN FRANCISCO 57 74 .435 16 B-ORG E-ORG O O O O
FRIDAY , AUGUST 30 SCHEDULE O O O O O
ATLANTA AT CHICAGO S-ORG O S-LOC
FLORIDA AT CINCINNATI S-ORG O S-LOC
SAN DIEGO AT MONTREAL B-ORG E-ORG O S-LOC
LOS ANGELES AT PHILADELPHIA B-ORG E-ORG O S-LOC
HOUSTON AT PITTSBURGH S-ORG O S-LOC
SAN FRANCISCO AT NEW YORK B-ORG E-ORG O B-LOC E-LOC
COLORADO AT ST LOUIS S-ORG O B-LOC E-LOC
-DOCSTART- O
BASEBALL - MAJOR LEAGUE RESULTS THURSDAY . O O B-MISC E-MISC O O O
NEW YORK 1996-08-30 B-LOC E-LOC O
Results of Major League O O B-MISC E-MISC
Baseball games played on Thursday ( home team in CAPS ) : O O O O O O O O O O O O
American League B-MISC E-MISC
DETROIT 4 Kansas City 1 S-ORG O B-ORG E-ORG O
Minnesota 6 MILWAUKEE 1 S-ORG O S-ORG O
CALIFORNIA 14 New York 3 S-ORG O B-ORG E-ORG O
SEATTLE 9 Baltimore 6 S-ORG O S-ORG O
National League B-MISC E-MISC
San Diego 3 NEW YORK 2 B-ORG E-ORG O B-ORG E-ORG O
Chicago 4 HOUSTON 3 S-ORG O S-ORG O
Cincinnati 18 COLORADO 7 S-ORG O S-ORG O
Atlanta 5 PITTSBURGH 1 S-ORG O S-ORG O
Los Angeles 2 MONTREAL 1 B-ORG E-ORG O S-ORG O
Florida 10 ST LOUIS 9 S-ORG O B-ORG E-ORG O
-DOCSTART- O
TENNIS - TARANGO , O'BRIEN SPRING TWIN UPSETS UNDER THE LIGHTS . O O S-PER O S-PER O O O O O O O
Larry Fine B-PER E-PER
NEW YORK 1996-08-30 B-LOC E-LOC O
Andre Agassi escaped disaster on Thursday but Wimbledon finalist MaliVai Washington and Marcelo Rios were not so fortunate on a night of upsets at the U.S. Open . B-PER E-PER O O O O O S-MISC O B-PER E-PER O B-PER E-PER O O O O O O O O O O O B-MISC E-MISC O
The 11th-seeded Washington fell short of reprising his Wimbledon miracle comeback as he lost to red-hot wildcard Alex O'Brien 6-3 6-4 5-7 3-6 6-3 in a two hour 51 minute struggle on the Stadium court . O O S-PER O O O O O S-MISC O O O O O O O O B-PER E-PER O O O O O O O O O O O O O O O O O
Next door on the grandstand , 10th seed Rios lost to another player with a Wimbledon connection -- bad boy Jeff Tarango . O O O O O O O O S-PER O O O O O O S-MISC O O O O B-PER E-PER O
The temperamental left-hander defeated the Chilean 6-4 4-6 7-6 6-2 . O O O O O S-MISC O O O O O
The day programme went smoothly although sixth-seeded former champion Agassi had to wriggle out of a dangerous 3-6 0-4 hole , winning 18 of the last 19 games against India 's Leander Paes . O O O O O O O O O S-PER O O O O O O O O O O O O O O O O O O O S-LOC O B-PER E-PER O
But the night belonged to the upstarts . O O O O O O O O
Washington , who climbed back from a 1-5 deficit , two sets down in the third set against Todd Martin in the Wimbledon semifinals , looked poised for another sensational comeback . S-PER O O O O O O O O O O O O O O O O O B-PER E-PER O O S-MISC O O O O O O O O O
O'Brien , a winner two weeks ago in New Haven for his first pro title , served for the match at 5-4 in the third set before Washington came charging back . S-PER O O O O O O O B-LOC E-LOC O O O O O O O O O O O O O O O O O S-PER O O O O
" I just kept saying to myself , ' keep giving yourself the best chance to win , keep battling , maybe something will happen , ' " said the 26-year-old O'Brien , ranked 65th . O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O S-PER O O O O
" I kept my composure and I was proud of myself for that -- usually I would have folded up the tent and gone home . " O O O O O O O O O O O O O O O O O O O O O O O O O O O
The hard-serving O'Brien , a former U.S. collegiate national champion , fired up 17 aces to ultimately subdue the never-say-die Washington . O O S-PER O O O S-LOC O O O O O O O O O O O O O S-PER O
The fifth set stayed on serve until the sixth game , when Washington , after saving one break point with a forehand winner down the line , netted a backhand to give O'Brien a 4-2 lead . O O O O O O O O O O O O S-PER O O O O O O O O O O O O O O O O O O O S-PER O O O O
The Texan blasted in two aces to hold serve at 5-2 and then converted his eighth match point for victory when Washington found the net with another backhand from 40-0 . O S-MISC O O O O O O O O O O O O O O O O O O O S-PER O O O O O O O O O
" You just kind of keep fighting and you keep trying to make him play a little bit . O O O O O O O O O O O O O O O O O O O
I think he got a little tight at a couple of moments , " said Washington . " O O O O O O O O O O O O O O O S-PER O O
But I think he served pretty well when he had to . " O O O O O O O O O O O O O
Tarango , whose Wimbledon tantrum two years ago brought him a $ 28,000 fine and suspension from this year 's tournament at the All-England Club , argued calls and taunted fans in his lively two hour , 24 minute tango with Rios on the grandstand . S-PER O O S-MISC O O O O O O O O O O O O O O O O O O O B-ORG E-ORG O O O O O O O O O O O O O O O O S-PER O O O O
A boisterous cheering section backed the distracted Chilean and booed the lanky American , who ate up all the attention . O O O O O O O S-MISC O O O O S-MISC O O O O O O O O
" I 'm an emotional player , " said the 104th-ranked Tarango . " O O O O O O O O O O O S-PER O O
I think I played very well tonight , very focused . " O O O O O O O O O O O O
The match turned on the third-set tiebreaker , which the American won 7-5 much to the dismay of the spectators . O O O O O O O O O O S-MISC O O O O O O O O O O
" I love the crowd if they boo me every day . O O O O O O O O O O O O
It fires me up , makes me play my best tennis , " Tarango said . O O O O O O O O O O O O O S-PER O O
" I played some of my best tennis in college when fraternities were throwing beer on me . O O O O O O O O O O O O O O O O O O
If tennis was like that every day , I think everybody wold be having a lot more fun . " O O O O O O O O O O O O O O O O O O O O
Rios did not appreciate Tarango 's antics . S-PER O O O S-PER O O O
" He 's always complaining too much , " said Rios . " O O O O O O O O O O S-PER O O
But I think it 's not that . O O O O O O O O
I think I played really bad . O O O O O O O
It was tough to play at night . O O O O O O O O
Balls were going really fast . O O O O O O
I lost too many points that I never lose . O O O O O O O O O O
I did n't play my tennis . " O O O O O O O O
" I do n't see the ball like I see during the day . O O O O O O O O O O O O O O
I play an American so that 's why I play at night . O O O S-MISC O O O O O O O O O
I did n't feel good on the court . " O O O O O O O O O O
At the end of the match , Tarango blew sarcastic kisses to the crowd , then jiggled his body to a Rios rooting section in a jeering salute . O O O O O O O S-PER O O O O O O O O O O O O O S-PER O O O O O O O
" I support their enthusiasm , " Tarango said about the fans . " O O O O O O O S-PER O O O O O O
At the same time , they 're cheering blatantly against me . O O O O O O O O O O O O
After I won I figured I could give them a little razzle-dazzle . " O O O O O O O O O O O O O O
-DOCSTART- O
NFL AMERICAN FOOTBALL-RANDALL CUNNINGHAM RETIRES . S-ORG B-MISC E-MISC S-PER O O
PHILADELPHIA 1996-08-29 S-LOC O
Randall Cunningham , the National Football League 's all-time leading rusher as a quarterback and one of the most athletic players ever to line up over centre , retired Thursday . B-PER E-PER O O B-ORG I-ORG E-ORG O O O O O O O O O O O O O O O O O O O O O O O O
Cunningham played his entire 11-year career with the Philadelphia Eagles . S-PER O O O O O O O B-ORG E-ORG O
A three-time Pro Bowl selection , Cunningham rushed for 4,482 yards on 677 carries . O O B-MISC E-MISC O O S-PER O O O O O O O O
" I would like to thank the Eagles organisation and the wonderful fans of Philadelphia for supporting me throughout my career , " Cunningham said . O O O O O O O S-ORG O O O O O O S-ORG O O O O O O O O S-PER O O
" Although it saddens me to leave , I am looking forward to spending more time with my family and pursuing other interests that have been on the back burner for sometime . " O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O
" Randall was one of the most exciting quarterbacks in NFL history , " said Eagles owner Jeffrey Lurie . " O S-PER O O O O O O O O S-ORG O O O O S-ORG O B-PER E-PER O O
During his 11 years in Philadelphia , Randall was the cornerstone of the Eagles ' franchise and brought many great moments to fans in Philadelphia as well as across the NFL . " O O O O O S-LOC O S-PER O O O O O S-ORG O O O O O O O O O O S-LOC O O O O O S-ORG O O
A second-round choice in 1985 , Cunningham completed 1,874-of-3,362 passes ( 55.7 percent ) for 22,877 yards and 150 touchdowns . O O O O O O S-PER O O O O O O O O O O O O O O
Cunningham has already been signed as a broadcaster . S-PER O O O O O O O O
-DOCSTART- O
GOLF - LEADING SCORES AT GREATER MILWAUKEE OPEN . O O O O O B-MISC I-MISC E-MISC O
MILWAUKEE , Wisconsin 1996-08-29 S-LOC O S-LOC O
Leading scores in O O O
the $ 1.2 million Greater Milwaukee Open at the par-71 , O O O O B-MISC I-MISC E-MISC O O O O
6,739-yard Brown Deer Park Golf Course after the first round O B-LOC I-LOC I-LOC I-LOC E-LOC O O O O
on Thursday ( players U.S. unless stated ) : O O O O S-LOC O O O O
62 Nolan Henke O B-PER E-PER
64 Bob Estes O B-PER E-PER
65 Billy Andrade , Duffy Waldorf , Jesper Parnevik ( Sweden ) O B-PER E-PER O B-PER E-PER O B-PER E-PER O S-LOC O
66 Neal Lancaster , Dave Barr ( Canada ) , Mike Sullivan , Willie O B-PER E-PER O B-PER E-PER O S-LOC O O B-PER E-PER O S-PER
Wood , Loren Roberts , Steve Stricker , Brian Claar , Russ Cochran S-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER
67 Mark Calcavecchia , Payne Stewart , Billy Mayfair , Ken O B-PER E-PER O B-PER E-PER O B-PER E-PER O S-PER
Green , Jerry Kelly , Tim Simpson , Olin Browne , Shane Bortsch , S-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Mike Hulbert , Brian Henninger , Tiger Woods , Steve Jurgenson , B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Bryan Gorman B-PER E-PER
-DOCSTART- O
GOLF - HENKE TAKES LEAD IN MILWAUKEE , WOODS MAKES PRO DEBUT . O O S-PER O O O S-LOC O S-PER O O O O
MILWAUKEE , Wisconsin 1996-08-29 S-LOC O S-LOC O
Nolan Henke fired a nine-under-par 62 to grab a two-shot lead after the opening round of the $ 1.2 million Greater Milwaukee Open Thursday as 20-year-old Tiger Woods shot 67 in his professional debut . B-PER E-PER O O O O O O O O O O O O O O O O O O B-MISC I-MISC E-MISC O O O B-PER E-PER O O O O O O O
Henke stood two strokes ahead of Bob Estes and three up on Billy Andrade , Duffy Waldorf and Jesper Parnevik . S-PER O O O O O B-PER E-PER O O O O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Woods , who turned pro Tuesday after winning an unprecedented third successive U.S. Amateur Championship , almost eagled the 18th hole . S-PER O O O O O O O O O O O B-MISC I-MISC E-MISC O O O O O O O
He settled for a birdie and a four-under opening round that left him five shots off the pace . O O O O O O O O O O O O O O O O O O O
" Yesterday was the toughest day I 've had for a long time , " Woods said . " O O O O O O O O O O O O O O O S-PER O O O
Today , I got to play golf . " O O O O O O O O O
He added : " I thought I got off off to a great start . O O O O O O O O O O O O O O O
It was a perfect start . O O O O O O
I 'm in a good position . " O O O O O O O O
Henke , who called his round a " pleasant surprise , " finished with six birdies on the final eight holes . S-PER O O O O O O O O O O O O O O O O O O O O O
" We finally got things going in the right direction , " he said . " O O O O O O O O O O O O O O O O
It was my best round in a very long time . O O O O O O O O O O O
My short game has improved since I 've had to use it so often . O O O O O O O O O O O O O O O
That 's always been the worst part of my game . O O O O O O O O O O O
All in all , playing bad 's been a good experience . " O O O O O O O O O O O O O
Henke , who came within one shot of the course record set by Andrew Magee during Wednesday 's pro-am , has three career PGA Tour victories , but none since the 1993 BellSouth Classic . S-PER O O O O O O O O O O O O B-PER E-PER O O O O O O O O B-MISC E-MISC O O O O O O O B-MISC E-MISC O
Estes , whose only win came at the 1994 Texas Open and whose best finish this year was a third-place tie at the Nortel Open in January , eagled the par-five fourth hole and added five birdies to grab sole possession of second place . S-PER O O O O O O O O B-MISC E-MISC O O O O O O O O O O O O B-MISC E-MISC O O O O O O O O O O O O O O O O O O O O
" No bogeys on the card , " he noted . " O O O O O O O O O O O O
Sometimes I take more pride in that . " O O O O O O O O O
Woods was among a group of 13 players at four under , including 1993 champion Billy Mayfair , who tied for second at last week 's World Series of Golf , and former U.S. Open champ Payne Stewart . S-PER O O O O O O O O O O O O O O B-PER E-PER O O O O O O O O O B-MISC I-MISC I-MISC E-MISC O O O B-MISC E-MISC O B-PER E-PER O
Defending champion Scott Hoch shot a three-under 68 and was six strokes back . O O B-PER E-PER O O O O O O O O O O
Phil Mickelson , the only four-time winner on the PGA Tour , skipped the tournament after winning the World Series of Golf last week . B-PER E-PER O O O O O O O B-MISC E-MISC O O O O O O O B-MISC I-MISC I-MISC E-MISC O O O
Mark Brooks , Tom Lehman and Mark O'Meara , who make up the rest of the top four on the money list , also took the week off . B-PER E-PER O B-PER E-PER O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O
-DOCSTART- O
SOCCER - SILVA 'S `LOST PASSPORT ' EXCUSE NOT ENOUGH FOR FIFA . O O S-PER O O O O O O O O S-ORG O
MADRID 1996-08-30 S-LOC O
Spanish first division team Deportivo Coruna will be without key midfielder Mauro Silva for Saturday 's game with Real Madrid after FIFA , soccer 's world governing body , suspended the Brazilian for one game for missing his national side 's European tour . S-MISC O O O B-ORG E-ORG O O O O O B-PER E-PER O O O O O B-ORG E-ORG O S-ORG O O O O O O O O O S-MISC O O O O O O O O O S-MISC O O
Silva excused his absence from Brazil 's game against Russia , on Wednesday , and Saturday 's match with the Netherlands by saying he had lost his passport . S-PER O O O O S-LOC O O O S-LOC O O O O O O O O O O S-LOC O O O O O O O O
But that did not prevent him from collecting the one-match suspension . O O O O O O O O O O O O
-DOCSTART- O
ATHLETICS - MITCHELL DEFEATS BAILEY IN FRONT OF FORMER CHAMPIONS . O O S-PER O S-PER O O O O O O
Adrian Warner B-PER E-PER
BERLIN 1996-08-30 S-LOC O
American Dennis Mitchell outclassed Olympic 100 metres champion Donovan Bailey for the third time at a major post-Games meeting in front of the most experienced sprinting crowd in the world on Friday . S-MISC B-PER E-PER O S-MISC O O O B-PER E-PER O O O O O O O S-MISC O O O O O O O O O O O O O O O
Watched by an array of former Olympic sprint champions at the Berlin grand prix meeting , Mitchell made a brilliant start in the 100 metres and held off Bailey 's strong finish to win in 10.08 seconds despite cool conditions . O O O O O O S-MISC O O O O S-LOC O O O O S-PER O O O O O O O O O O O S-PER O O O O O O O O O O O O
Bailey , who set a world record of 9.84 on his way to victory in Atlanta , could not catch his American rival and had to settle for third in a tight finish . S-PER O O O O O O O O O O O O O O S-LOC O O O O O S-MISC O O O O O O O O O O O O
Jamaica 's Michael Green was second with 10.09 with Bailey finishing in 10.13 . S-LOC O B-PER E-PER O O O O O S-PER O O O O
Last Friday Mitchell , who finished fourth at the Atlanta Games , upstaged a trio of Olympic champions including Bailey to win the 100 in Brussels . O O S-PER O O O O O O B-MISC E-MISC O O O O O S-MISC O O S-PER O O O O O S-LOC O
Earlier this month he also beat world champion Bailey in Zurich . O O O O O O O O S-PER O S-LOC O
Berlin , Brussels and Zurich all belong to the most lucrative series in the sport , the Golden Four . S-LOC O S-LOC O S-LOC O O O O O O O O O O O O B-MISC E-MISC O
Among the crowd on Friday were Olympic 100 metres champions going back to 1948 . O O O O O O S-MISC O O O O O O O O
They had been invited to the meeting to watch a special relay to mark the 60th anniversary of Jesse Owens 's four gold medals at the 1936 Olympics in the same Berlin stadium . O O O O O O O O O O O O O O O O O O B-PER E-PER O O O O O O O S-MISC O O O S-LOC O O
" Today the concentration was the most important thing for me , " Mitchell said . O O O O O O O O O O O O O S-PER O O
Despite the coolish conditions American Olympic champion Gail Devers looked in commanding form in the women 's 100 , clocking 10.89 to defeat Jamaican rival Merlene Ottey , who was second in 10.94 . O O O O B-MISC E-MISC O B-PER E-PER O O O O O O O O O O O O O O S-MISC O B-PER E-PER O O O O O O O
-DOCSTART- O
ATHLETICS - BERLIN GRAND PRIX RESULTS . O O B-MISC I-MISC E-MISC O O
BERLIN 1996-08-30 S-LOC O
Leading results at the Berlin O O O O S-MISC
Grand Prix athletics meeting on Friday : B-MISC E-MISC O O O O O
Women 's 100 metres hurdles O O O O O
1. Michelle Freeman ( Jamaica ) 12.71 seconds O B-PER E-PER O S-LOC O O O
2. Ludmila Engquist ( Sweden ) 12.74 O B-PER E-PER O S-LOC O O
3. Aliuska Lopez ( Cuba ) 12.92 O B-PER E-PER O S-LOC O O
4. Brigita Bokovec ( Slovenia ) 12.92 O B-PER E-PER O S-LOC O O
5. Dionne Rose ( Jamaica ) 12.92 O B-PER E-PER O S-LOC O O
6. Julie Baumann ( Switzerland ) 13.11 O B-PER E-PER O S-LOC O O
7. Gillian Russell ( Jamaica ) 13.17 O B-PER E-PER O S-LOC O O
Women 's 1,500 metres O O O O
1. Svetlana Masterkova ( Russia ) four minutes 6.87 seconds O B-PER E-PER O S-LOC O O O O O
2. Patricia Djate-Taillard ( France ) 4:08.22 O B-PER E-PER O S-LOC O O
3. Carla Sacramento ( Portugal ) 4:08.96 O B-PER E-PER O S-LOC O O
4. Yekaterina Podkopayeva ( Russia ) 4:09.25 O B-PER E-PER O S-LOC O O
5. Leah Pells ( Canada ) 4:09.95 O B-PER E-PER O S-LOC O O
6. Carmen Wuestenhagen ( Germany ) 4:10.38 O B-PER E-PER O S-LOC O O
7. Margarita Maruseva ( Russia ) 4:10.87 O B-PER E-PER O S-LOC O O
8. Sara Thorsett ( U.S. ) 4:11.06 O B-PER E-PER O S-LOC O O
Men 's 110 metres hurdles O O O O O
1. Mark Crear ( U.S. ) 13.26 seconds O B-PER E-PER O S-LOC O O O
2. Tony Jarrett ( Britain ) 13.35 O B-PER E-PER O S-LOC O O
3. Florian Schwarthoff ( Germany ) 13.36 O B-PER E-PER O S-LOC O O
4. Emilio Valle ( Cuba ) 13.52 O B-PER E-PER O S-LOC O O
5. Falk Balzer ( Germany ) 13.52 O B-PER E-PER O S-LOC O O
6. Steve Brown ( U.S. ) 13.53 O B-PER E-PER O S-LOC O O
7. Frank Busemann ( Germany ) 13.58 O B-PER E-PER O S-LOC O O
8. Jack Pierce ( U.S. ) 13.60 O B-PER E-PER O S-LOC O O
Men 's 200 metres O O O O
1. Frankie Fredericks ( Namibia ) 19.97 seconds O B-PER E-PER O S-LOC O O O
2. Michael Johnson ( U.S. ) 20.02 O B-PER E-PER O S-LOC O O
3. Ato Boldon ( Trinidad ) 20.37 O B-PER E-PER O S-LOC O O
4. Geir Moen ( Norway ) 20.41 O B-PER E-PER O S-LOC O O
5. Patrick Stevens ( Belgium ) 20.54 O B-PER E-PER O S-LOC O O
6. Jon Drummond ( U.S. ) 20.78 O B-PER E-PER O S-LOC O O
7. Claus Hirsbro ( Denmark ) 20.90 O B-PER E-PER O S-LOC O O
8. Ivan Garcia ( Cuba ) 20.96 O B-PER E-PER O S-LOC O O
Women 's shot put O O O O
1. Astrid Kumbernuss ( Germany ) 19.89 metres O B-PER E-PER O S-LOC O O O
2. Claudia Mues ( Germany ) 18.80 O B-PER E-PER O S-LOC O O
3. Irina Korzhanenko ( Russia ) 18.63 O B-PER E-PER O S-LOC O O
4. Valentina Fedyushina ( Russia ) 18.55 O B-PER E-PER O S-LOC O O
5. Stephanie Storp ( Germany ) 18.41 O B-PER E-PER O S-LOC O O
Men 's mile O O O
1. Noureddine Morceli ( Algeria ) 3 minutes 49.09 seconds O B-PER E-PER O S-LOC O O O O O
2. Venuste Niyongabo ( Burundi ) 3:51.01 O B-PER E-PER O S-LOC O O
3. William Tanui ( Kenya ) 3:51.40 O B-PER E-PER O S-LOC O O
4. Laban Rotich ( Kenya ) 3:53.42 O B-PER E-PER O S-LOC O O
5. Marko Koers ( Netherlands ) 3:53.47 O B-PER E-PER O S-LOC O O
6. Isaac Viciosa ( Spain ) 3:53.85 O B-PER E-PER O S-LOC O O
7. John Mayock ( Britain ) 3:54.67 O B-PER E-PER O S-LOC O O
8. Marcus O'Sullivan ( Ireland ) 3:54.87 O B-PER E-PER O S-LOC O O
Men 's discus O O O
1. Lars Riedel ( Germany ) 70.60 metres O B-PER E-PER O S-LOC O O O
2. Anthony Washington ( U.S. ) 68.44 O B-PER E-PER O S-LOC O O
3. Vasily Kaptyukh ( Belarus ) 66.24 O B-PER E-PER O S-LOC O O
4. Vladimir Dubrovshchik ( Belarus ) 65.30 O B-PER E-PER O S-LOC O O
5. Virgilijus Alekna ( Lithuania ) 65.00 O B-PER E-PER O S-LOC O O
6. Juergen Schult ( Germany ) 64.46 O B-PER E-PER O S-LOC O O
7. Andreas Seelig ( Germany ) 62.00 O B-PER E-PER O S-LOC O O
8. Michael Moellenbeck ( Germany ) 58.56 O B-PER E-PER O S-LOC O O
Women 's 100 metres O O O O
1. Gail Devers ( U.S. ) 10.89 seconds O B-PER E-PER O S-LOC O O O
2. Merlene Ottey ( Jamaica ) 10.94 O B-PER E-PER O S-LOC O O
3. Gwen Torrence ( U.S. ) 11.07 O B-PER E-PER O S-LOC O O
4. Mary Onyali ( Nigeria ) 11.14 O B-PER E-PER O S-LOC O O
5. Chryste Gaines ( U.S. ) 11.20 O B-PER E-PER O S-LOC O O
6. Chandra Sturrup ( Bahamas ) 11.26 O B-PER E-PER O S-LOC O O
7. Irina Privalova ( Russia ) 11.27 O B-PER E-PER O S-LOC O O
8. Inger Miller ( U.S. ) 11.37 O B-PER E-PER O S-LOC O O
Women 's 5,000 metres O O O O
1. Gabriela Szabo ( Romania ) 15 minutes 04.95 seconds O B-PER E-PER O S-LOC O O O O O
2. Gete Wami ( Ethiopia ) 15:05.21 O B-PER E-PER O S-LOC O O
3. Rose Cheruiyot ( Kenya ) 15:05.41 O B-PER E-PER O S-LOC O O
4. Annemari Sandell ( Finland ) 15:06.33 O B-PER E-PER O S-LOC O O
5. Tegla Loroupe ( Kenya ) 15:08.79 O B-PER E-PER O S-LOC O O
6. Gunhild Halle ( Norway ) 15:09.00 O B-PER E-PER O S-LOC O O
7. Pauline Konga ( Kenya ) 15:09.74 O B-PER E-PER O S-LOC O O
8. Sally Barsosio ( Kenya ) 15:14.34 O B-PER E-PER O S-LOC O O
Men 's 400 metres hurdles O O O O O
1. Torrance Zellner ( U.S. ) 48.23 seconds O B-PER E-PER O S-LOC O O O
2. Samuel Matete ( Zambia ) 48.34 O B-PER E-PER O S-LOC O O
3. Derrick Adkins ( U.S. ) 48.62 O B-PER E-PER O S-LOC O O
4. Fabrizio Mori ( Italy ) 49.21 O B-PER E-PER O S-LOC O O
5. Sven Nylander ( Sweden ) 49.22 O B-PER E-PER O S-LOC O O
6. Eric Thomas ( U.S. ) 49.35 O B-PER E-PER O S-LOC O O
Не удается отобразить этот файл, потому что он содержит неожиданный символ в строке 87 и столбце 81.

Просмотреть файл

@ -0,0 +1,500 @@
-DOCSTART- O
SOCCER - JAPAN GET LUCKY WIN , CHINA IN SURPRISE DEFEAT . O O S-LOC O O O O S-PER O O O O
Nadim Ladki B-PER E-PER
AL-AIN , United Arab Emirates 1996-12-06 S-LOC O B-LOC I-LOC E-LOC O
Japan began the defence of their Asian Cup title with a lucky 2-1 win against Syria in a Group C championship match on Friday . S-LOC O O O O O B-MISC E-MISC O O O O O O O S-LOC O O O O O O O O O
But China saw their luck desert them in the second match of the group , crashing to a surprise 2-0 defeat to newcomers Uzbekistan . O S-LOC O O O O O O O O O O O O O O O O O O O O O S-LOC O
China controlled most of the match and saw several chances missed until the 78th minute when Uzbek striker Igor Shkvyrin took advantage of a misdirected defensive header to lob the ball over the advancing Chinese keeper and into an empty net . S-LOC O O O O O O O O O O O O O O O S-MISC O B-PER E-PER O O O O O O O O O O O O O O S-MISC O O O O O O O
Oleg Shatskiku made sure of the win in injury time , hitting an unstoppable left foot shot from just outside the area . B-PER E-PER O O O O O O O O O O O O O O O O O O O O O
The former Soviet republic was playing in an Asian Cup finals tie for the first time . O O S-MISC O O O O O B-MISC E-MISC O O O O O O O
Despite winning the Asian Games title two years ago , Uzbekistan are in the finals as outsiders . O O O B-MISC E-MISC O O O O O S-LOC O O O O O O O
Two goals from defensive errors in the last six minutes allowed Japan to come from behind and collect all three points from their opening meeting against Syria . O O O O O O O O O O O S-LOC O O O O O O O O O O O O O O S-LOC O
Takuya Takagi scored the winner in the 88th minute , rising to head a Hiroshige Yanagimoto cross towards the Syrian goal which goalkeeper Salem Bitar appeared to have covered but then allowed to slip into the net . B-PER E-PER O O O O O O O O O O O O B-PER E-PER O O O S-MISC O O O B-PER E-PER O O O O O O O O O O O O O
It was the second costly blunder by Syria in four minutes . O O O O O O O S-LOC O O O O
Defender Hassan Abbas rose to intercept a long ball into the area in the 84th minute but only managed to divert it into the top corner of Bitar 's goal . O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O O O O S-PER O O O
Nader Jokhadar had given Syria the lead with a well-struck header in the seventh minute . B-PER E-PER O O S-LOC O O O O O O O O O O O
Japan then laid siege to the Syrian penalty area for most of the game but rarely breached the Syrian defence . S-LOC O O O O O S-MISC O O O O O O O O O O O S-MISC O O
Bitar pulled off fine saves whenever they did . S-PER O O O O O O O O
Japan coach Shu Kamo said : ' ' The Syrian own goal proved lucky for us . S-LOC O B-PER E-PER O O O O O S-MISC O O O O O O O
The Syrians scored early and then played defensively and adopted long balls which made it hard for us . ' O S-MISC O O O O O O O O O O O O O O O O O O
' O
Japan , co-hosts of the World Cup in 2002 and ranked 20th in the world by FIFA , are favourites to regain their title here . S-LOC O O O O B-MISC E-MISC O O O O O O O O O S-ORG O O O O O O O O O
Hosts UAE play Kuwait and South Korea take on Indonesia on Saturday in Group A matches . O S-LOC O S-LOC O B-LOC E-LOC O O S-LOC O O O O O O O
All four teams are level with one point each from one game . O O O O O O O O O O O O O
-DOCSTART- O
RUGBY UNION - CUTTITTA BACK FOR ITALY AFTER A YEAR . B-ORG E-ORG O S-PER O O S-LOC O O O O
ROME 1996-12-06 S-LOC O
Italy recalled Marcello Cuttitta S-LOC O B-PER E-PER
on Friday for their friendly against Scotland at Murrayfield more than a year after the 30-year-old wing announced he was retiring following differences over selection . O O O O O O S-LOC O S-LOC O O O O O O O O O O O O O O O O O
Cuttitta , who trainer George Coste said was certain to play on Saturday week , was named in a 21-man squad lacking only two of the team beaten 54-21 by England at Twickenham last month . S-PER O O O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O O O O S-LOC O S-LOC O O O
Stefano Bordon is out through illness and Coste said he had dropped back row Corrado Covi , who had been recalled for the England game after five years out of the national team . B-PER E-PER O O O O O S-PER O O O O O O B-PER E-PER O O O O O O O S-LOC O O O O O O O O O O
Cuttitta announced his retirement after the 1995 World Cup , where he took issue with being dropped from the Italy side that faced England in the pool stages . S-PER O O O O O B-MISC I-MISC E-MISC O O O O O O O O O O S-LOC O O O S-LOC O O O O O
Coste said he had approached the player two months ago about a comeback . S-PER O O O O O O O O O O O O O
" He ended the World Cup on the wrong note , " Coste said . O O O O B-MISC E-MISC O O O O O O S-PER O O
" I thought it would be useful to have him back and he said he would be available . O O O O O O O O O O O O O O O O O O O
I think now is the right time for him to return . " O O O O O O O O O O O O O
Squad : Javier Pertile , Paolo Vaccari , Marcello Cuttitta , Ivan Francescato , Leandro Manteri , Diego Dominguez , Francesco Mazzariol , Alessandro Troncon , Orazio Arancio , Andrea Sgorlon , Massimo Giovanelli , Carlo Checchinato , Walter Cristofoletto , Franco Properzi Curti , Carlo Orlandi , Massimo Cuttitta , Giambatista Croci , Gianluca Guidi , Nicola Mazzucato , Alessandro Moscardi , Andrea Castellani . O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER I-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
-DOCSTART- O
SOCCER - LATE GOALS GIVE JAPAN WIN OVER SYRIA . O O O O O S-LOC O O S-LOC O
AL-AIN , United Arab Emirates 1996-12-06 S-LOC O B-LOC I-LOC E-LOC O
Two goals in the last six minutes gave holders Japan an uninspiring 2-1 Asian Cup victory over Syria on Friday . O O O O O O O O O S-LOC O O O B-MISC E-MISC O O S-LOC O O O
Takuya Takagi headed the winner in the 88th minute of the group C game after goalkeeper Salem Bitar spoiled a mistake-free display by allowing the ball to slip under his body . B-PER E-PER O O O O O O O O O O O O O O B-PER E-PER O O O O O O O O O O O O O O
It was the second Syrian defensive blunder in four minutes . O O O O S-MISC O O O O O O
Defender Hassan Abbas rose to intercept a long ball into the area in the 84th minute but only managed to divert it into the top corner of Bitar 's goal . O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O O O O S-PER O O O
Syria had taken the lead from their first serious attack in the seventh minute . S-LOC O O O O O O O O O O O O O O
Nader Jokhadar headed a cross from the right by Ammar Awad into the top right corner of Kenichi Shimokawa 's goal . B-PER E-PER O O O O O O O B-PER E-PER O O O O O O B-PER E-PER O O O
Japan then laid siege to the Syrian penalty area and had a goal disallowed for offside in the 16th minute . S-LOC O O O O O S-MISC O O O O O O O O O O O O O O
A minute later , Bitar produced a good double save , first from Kazuyoshi Miura 's header and then blocked a Takagi follow-up shot . O O O O S-PER O O O O O O O O B-PER E-PER O O O O O O S-PER O O O
Bitar saved well again from Miura in the 37th minute , parrying away his header from a corner . S-PER O O O O S-PER O O O O O O O O O O O O O
Japan started the second half brightly but Bitar denied them an equaliser when he dived to his right to save Naoki Soma 's low drive in the 53rd minute . S-LOC O O O O O O S-PER O O O O O O O O O O O O B-PER E-PER O O O O O O O O
Japan : 19 - Kenichi Shimokawa , 2 - Hiroshige Yanagimoto , 3 - Naoki Soma , 4 - Masami Ihara , 5 - Norio Omura , 6 - Motohiro Yamaguchi , 8 - Masakiyo Maezono ( 7 - Yasuto Honda 71 ) , 9 - Takuya Takagi , 10 - Hiroshi Nanami , 11 - Kazuyoshi Miura , 15 - Hiroaki Morishima ( 14 - Masayuki Okano 75 ) . S-LOC O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O
Syria : 24 - Salem Bitar , 3 - Bachar Srour ; 4 - Hassan Abbas , 5 - Tarek Jabban , 6 - Ammar Awad ( 9 - Louay Taleb 69 ) , 8 - Nihad al-Boushi , 10 - Mohammed Afash , 12 - Ali Dib , 13 - Abdul Latif Helou ( 17 - Ammar Rihawiy 46 ) , 14 - Khaled Zaher ; 16 - Nader Jokhadar . S-LOC O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER I-PER E-PER O O O B-PER E-PER O O O O O B-PER E-PER O O O B-PER E-PER O
-DOCSTART- O
FREESTYLE SKIING-WORLD CUP MOGUL RESULTS . O B-MISC E-MISC O O O
TIGNES , France 1996-12-06 S-LOC O S-LOC O
Results of the World Cup O O O B-MISC E-MISC
freestyle skiing moguls competition on Friday : O O O O O O O
Men O
1. Jesper Ronnback ( Sweden ) 25.76 points O B-PER E-PER O S-LOC O O O
2. Andrei Ivanov ( Russia ) 24.88 O B-PER E-PER O S-LOC O O
3. Ryan Johnson ( Canada ) 24.57 O B-PER E-PER O S-LOC O O
4. Jean-Luc Brassard ( Canada ) 24.40 O B-PER E-PER O S-LOC O O
5. Korneilus Hole ( Norway ) 23.92 O B-PER E-PER O S-LOC O O
6. Jeremie Collomb-Patton ( France ) 23.87 O B-PER E-PER O S-LOC O O
7. Jim Moran ( U.S. ) 23.25 O B-PER E-PER O S-LOC O O
8. Dominick Gauthier ( Canada ) 22.73 O B-PER E-PER O S-LOC O O
9. Johann Gregoire ( France ) 22.58 O B-PER E-PER O S-LOC O O
10. Troy Benson ( U.S. ) 22.56 O B-PER E-PER O S-LOC O O
Women O
1. Tatjana Mittermayer ( Germany ) 24.32 O B-PER E-PER O S-LOC O O
2. Candice Gilg ( France ) 24.31 O B-PER E-PER O S-LOC O O
3. Minna Karhu ( Finland ) 24.05 O B-PER E-PER O S-LOC O O
4. Tae Satoya ( Japan ) 23.75 O B-PER E-PER O S-LOC O O
5. Ann Battellle ( U.S. ) 23.56 O B-PER E-PER O S-LOC O O
6. Donna Weinbrecht ( U.S. ) 22.48 O B-PER E-PER O S-LOC O O
7. Liz McIntyre ( U.S. ) 22.00 O B-PER E-PER O S-LOC O O
8. Elena Koroleva ( Russia ) 21.77 O B-PER E-PER O S-LOC O O
9. Ljudmila Dymchenko ( Russia ) 21.59 O B-PER E-PER O S-LOC O O
10. Katleen Allais ( France ) 21.58 O B-PER E-PER O S-LOC O O
-DOCSTART- O
SOCCER - ASIAN CUP GROUP C RESULTS . O O B-MISC E-MISC O O O O
AL-AIN , United Arab Emirates 1996-12-06 S-LOC O B-LOC I-LOC E-LOC O
Results of Asian Cup group C matches played on Friday : O O B-MISC E-MISC O O O O O O O
Japan 2 Syria 1 ( halftime 0-1 ) S-LOC O S-LOC O O O O O
Scorers : O O
Japan - Hassan Abbas 84 own goal , Takuya Takagi 88 . S-LOC O B-PER E-PER O O O O B-PER E-PER O O
Syria - Nader Jokhadar 7 S-LOC O B-PER E-PER O
Attendance : 10,000 . O O O O
China 0 Uzbekistan 2 ( halftime 0-0 ) S-LOC O S-LOC O O O O O
Scorers : Shkvyrin Igor 78 , Shatskikh Oleg 90 O O B-PER E-PER O O B-PER E-PER O
Attendence : 3,000 O O O
Standings ( tabulate under played , won , drawn , lost , goals O O O O O O O O O O O O O
for , goals against , points ) : O O O O O O O O
Uzbekistan 1 1 0 0 2 0 3 S-LOC O O O O O O O
Japan 1 1 0 0 2 1 3 S-LOC O O O O O O O
Syria 1 0 0 1 1 2 0 S-LOC O O O O O O O
China 1 0 0 1 0 2 0 S-LOC O O O O O O O
-DOCSTART- O
CRICKET - PAKISTAN V NEW ZEALAND ONE-DAY SCOREBOARD . O O S-LOC O B-LOC E-LOC O O O
[ CORRECTED 14:06 GMT ] O O O S-MISC O
SIALKOT , Pakistan 1996-12-06 S-LOC O S-LOC O
Scoreboard in the second O O O O
one-day cricket international between Pakistan and New Zealand O O O O S-LOC O B-LOC E-LOC
on Friday : O O O
Pakistan S-LOC
Saeed Anwar run out 91 ( corrects from 90 ) B-PER E-PER O O O O O O O O
Zahoor Elahi b Cairns 86 ( corrects from 87 ) B-PER E-PER O S-PER O O O O O O
Ijaz Ahmad c Spearman b Vaughan 59 B-PER E-PER O S-PER O S-PER O
Inzamamul Haq st Germon b Astle 2 B-PER E-PER O S-PER O S-PER O
Wasim Akram b Harris 4 B-PER E-PER O S-PER O
Shahid Afridi b Harris 2 B-PER E-PER O S-PER O
Moin Khan c Astle b Harris 1 B-PER E-PER O S-PER O S-PER O
Waqar Younis st Germon b Harris 0 B-PER E-PER O S-PER O S-PER O
Saqlain Mushtaq b Harris 2 B-PER E-PER O S-PER O
Mushtaq Ahmad not out 5 B-PER E-PER O O O
Salim Malik not out 1 B-PER E-PER O O O
Extras ( lb-8 nb-2 w-14 ) 24 O O O O O O O
Total ( for 9 wickets in 47 overs ) 277 O O O O O O O O O O
Fall of wicket : 1-177 ( corrects from 1-178 ) 2-225 3-240 4-247 5-252 6-260 7-261 8-269 9-276 O O O O O O O O O O O O O O O O O O
Bowling : Doull 8-1-60-0 ( w-3 ) , Kennedy 3-0-24-0 ( w-7 nb-1 ) , O O S-PER O O O O O S-PER O O O O O O
Cairns 8-1-35-1 ( w-2 ) , Vaughan 9-1-55-1 , Harris 10-0-42-5 ( w-1 ) , S-PER O O O O O S-PER O O S-PER O O O O O
Astle 9-0-53-1 ( w-1 nb-1 ) S-PER O O O O O
New Zealand innings B-LOC E-LOC O
B. Young c Moin Khan b Waqar 5 B-PER E-PER O B-PER E-PER O S-PER O
C. Spearman c Moin Khan b Wasim 0 B-PER E-PER O B-PER E-PER O S-PER O
A. Parore c Ijaz Ahmad b Saqlain 37 B-PER E-PER O B-PER E-PER O S-PER O
S. Fleming c and b Afridi 88 B-PER E-PER O O O S-PER O
C. Cairns b Saqlain 10 B-PER E-PER O S-PER O
N. Astle c Ijaz Ahmad b Salim Malik 20 B-PER E-PER O B-PER E-PER O B-PER E-PER O
C. Harris lbw b Wasim 22 B-PER E-PER O O S-PER O
L. Germon lbw b Afridi 2 B-PER E-PER O O S-PER O
J. Vaughan c Moin Khan b Wasim 13 B-PER E-PER O B-PER E-PER O S-PER O
S. Doull c subs ( M. Wasim ) b Waqar 1 B-PER E-PER O O O B-PER E-PER O O S-PER O
R. Kennedy not out 7 B-PER E-PER O O O
Extras ( b-9 lb-3 w-12 nb-2 ) 26 O O O O O O O O
Total ( all out in 42.1 overs ) 231 O O O O O O O O O
Fall of wickets : 1-3 2-7 3-125 4-146 5-170 6-190 7-195 O O O O O O O O O O O
8-213 9-216 . O O O
Bowling : Wasim Akram 8.1-0-43-3 ( 9w , 1nb ) , Waqar Younis O O B-PER E-PER O O O O O O O B-PER E-PER
6-0-32-2 ( 2w , 1nb ) , Saqlain Mushtaq 8-0-54-2 , Mushtaq Ahmad O O O O O O O B-PER E-PER O O B-PER E-PER
10-0-42-0 ( 1w ) , Shahid Afridi 7-0-40-2 , Salim Malik 2.5-0-8-1 , O O O O O B-PER E-PER O O B-PER E-PER O O
Ijaz Ahmad 0.1-0-0-0 . B-PER E-PER O O
Result : Pakistan won by 46 runs . O O S-LOC O O O O O
Third one-day match : December 8 , in Karachi . O O O O O O O O S-LOC O
-DOCSTART- O
SOCCER - ENGLISH F.A. CUP SECOND ROUND RESULT . O O B-MISC I-MISC E-MISC O O O O
LONDON 1996-12-06 S-LOC O
Result of an English F.A. Challenge O O O B-MISC I-MISC E-MISC
Cup second round match on Friday : S-MISC O O O O O O
Plymouth 4 Exeter 1 S-ORG O S-ORG O
-DOCSTART- O
SOCCER - BLINKER BAN LIFTED . O O S-PER O O O
LONDON 1996-12-06 S-LOC O
Dutch forward Reggie Blinker had his indefinite suspension lifted by FIFA on Friday and was set to make his Sheffield Wednesday comeback against Liverpool on Saturday . S-MISC O B-PER E-PER O O O O O O S-ORG O O O O O O O O B-ORG E-ORG O O S-ORG O O O
Blinker missed his club 's last two games after FIFA slapped a worldwide ban on him for appearing to sign contracts for both Wednesday and Udinese while he was playing for Feyenoord . S-PER O O O O O O O O S-ORG O O O O O O O O O O O O O S-ORG O S-ORG O O O O O S-ORG O
FIFA 's players ' status committee , meeting in Barcelona , decided that although the Udinese document was basically valid , it could not be legally protected . S-ORG O O O O O O O O S-LOC O O O O O S-ORG O O O O O O O O O O O O
The committee said the Italian club had violated regulations by failing to inform Feyenoord , with whom the player was under contract . O O O O S-MISC O O O O O O O O S-ORG O O O O O O O O O
Blinker was fined 75,000 Swiss francs ( $ 57,600 ) for failing to inform the Engllsh club of his previous commitment to Udinese . S-PER O O O S-MISC O O O O O O O O O O S-MISC O O O O O O S-ORG O
-DOCSTART- O
SOCCER - LEEDS ' BOWYER FINED FOR PART IN FAST-FOOD FRACAS . O O S-ORG O S-PER O O O O O O O
LONDON 1996-12-06 S-LOC O
Leeds ' England under-21 striker Lee Bowyer was fined 4,500 pounds ( $ 7,400 ) on Friday for hurling chairs at restaurant staff during a disturbance at a McDonald 's fast-food restaurant . S-ORG O S-LOC O O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O B-ORG E-ORG O O O
Bowyer , 19 , who was caught in the act by security cameras , pleaded guilty to a charge of affray at a court in London . S-PER O O O O O O O O O O O O O O O O O O O O O O O O S-LOC O
He was fined and ordered to pay a total of 175 pounds to two members of staff injured in the fracas in an east London restaurant in October . O O O O O O O O O O O O O O O O O O O O O O O O S-LOC O O O O
Leeds had already fined Bowyer 4,000 pounds ( $ 6,600 ) and warned him a repeat of his criminal behaviour could cost him his place in the side . S-ORG O O O S-PER O O O O O O O O O O O O O O O O O O O O O O O O
Bowyer , who moved to the Yorkshire club in August for 3.5 million pounds ( $ 5.8 million ) , was expected to play against Middlesbrough on Saturday . S-PER O O O O O S-LOC O O O O O O O O O O O O O O O O O O S-ORG O O O
-DOCSTART- O
BASKETBALL - EUROLEAGUE STANDINGS . O O S-MISC O O
LONDON 1996-12-06 S-LOC O
Standings in the men 's EuroLeague O O O O O S-MISC
basketball championship after Thursday 's matches ( tabulate under O O O O O O O O O
played , won , lost , points ) : O O O O O O O O O
Group A O O
CSKA Moscow ( Russia 9 6 3 15 B-ORG E-ORG O S-LOC O O O O
Stefanel Milan ( Italy ) 9 6 3 15 B-ORG E-ORG O S-LOC O O O O O
Maccabi Tel Aviv ( Israel ) 9 5 4 14 B-ORG I-ORG E-ORG O S-LOC O O O O O
Ulker Spor ( Turkey ) 9 4 5 13 B-ORG E-ORG O S-LOC O O O O O
Limoges ( France ) 9 3 6 12 S-ORG O S-LOC O O O O O
Panionios ( Greece ) 9 3 6 12 S-ORG O S-LOC O O O O O
Group B O O
Teamsystem Bologna ( Italy ) 9 7 2 16 B-ORG E-ORG O S-LOC O O O O O
Olympiakos ( Greece ) 9 5 4 14 S-ORG O S-LOC O O O O O
Cibona Zagreb ( Croatia ) 9 5 4 14 B-ORG E-ORG O S-LOC O O O O O
Alba Berlin ( Germany ) 9 5 4 14 B-ORG E-ORG O S-LOC O O O O O
Estudiantes Madrid ( Spain ) 9 5 4 14 B-ORG E-ORG O S-LOC O O O O O
Charleroi ( Belgium ) 9 0 9 9 S-ORG O S-LOC O O O O O
Group C O O
Panathinaikos ( Greece ) 9 7 2 16 S-ORG O S-LOC O O O O O
Ljubljana ( Slovenia ) 9 6 3 15 S-ORG O S-LOC O O O O O
Villeurbanne ( France ) 9 6 3 15 S-ORG O S-LOC O O O O O
Barcelona ( Spain ) 9 4 5 13 S-ORG O S-LOC O O O O O
Split ( Croatia ) 9 4 5 13 S-ORG O S-LOC O O O O O
Bayer Leverkusen ( Germany ) 9 0 9 9 B-ORG E-ORG O S-LOC O O O O O
Group D O O
Efes Pilsen ( Turkey ) 9 7 2 16 B-ORG E-ORG O S-LOC O O O O O
Pau-Orthez ( France ) 9 5 4 14 S-ORG O S-LOC O O O O O
Partizan Belgrade ( Yugoslavia ) 9 5 4 14 B-ORG E-ORG O S-LOC O O O O O
Kinder Bologna ( Italy ) 9 4 5 13 B-ORG E-ORG O S-LOC O O O O O
Sevilla ( Spain ) 9 4 5 13 S-ORG O S-LOC O O O O O
Dynamo Moscow ( Russia ) 9 2 7 11 B-ORG E-ORG O S-LOC O O O O O
-DOCSTART- O
RUGBY UNION - LITTLE TO MISS CAMPESE FAREWELL . B-ORG E-ORG O S-PER O O S-PER O O
Robert Kitson B-PER E-PER
LONDON 1996-12-06 S-LOC O
Centre Jason Little will miss Australia 's end-of-tour fixture against the Barbarians at Twickenham on Saturday . O B-PER E-PER O O S-LOC O O O O O S-ORG O S-LOC O O O
Little has opted not to risk aggravating the knee injury which ruled him out of a large chunk of the tour and is replaced by fellow Queenslander Daniel Herbert . S-PER O O O O O O O O O O O O O O O O O O O O O O O O O S-MISC B-PER E-PER O
Owen Finegan has recovered from the knocks he took in last weekend 's test against Wales and retains his place in the back-row ahead of Daniel Manu . B-PER E-PER O O O O O O O O O O O O O S-LOC O O O O O O O O O B-PER E-PER O
The Wallabies have their sights set on a 13th successive victory to end their European tour with a 100 percent record but also want to turn on the style and provide David Campese with a fitting send-off in his final match in Australian colours . O S-ORG O O O O O O O O O O O O S-MISC O O O O O O O O O O O O O O O O B-PER E-PER O O O O O O O O O S-MISC O O
The Wallabies currently have no plans to make any special presentation to the 34-year-old winger but a full house of 75,000 spectators will still gather in the hope of witnessing one last moment of magic . O S-ORG O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O O
Campese will be up against a familiar foe in the shape of Barbarians captain Rob Andrew , the man who kicked Australia to defeat with a last-ditch drop-goal in the World Cup quarter-final in Cape Town . S-PER O O O O O O O O O O O S-ORG O B-PER E-PER O O O O O S-LOC O O O O O O O O B-MISC E-MISC O O B-LOC E-LOC O
" Campo has a massive following in this country and has had the public with him ever since he first played here in 1984 , " said Andrew , also likely to be making his final Twickenham appearance . O S-PER O O O O O O O O O O O O O O O O O O O O O O O O O S-PER O O O O O O O O S-LOC O O
On tour , Australia have won all four tests against Italy , Scotland , Ireland and Wales , and scored 414 points at an average of almost 35 points a game . O O O S-LOC O O O O O O S-LOC O S-LOC O S-LOC O S-LOC O O O O O O O O O O O O O O O
League duties restricted the Barbarians ' selectorial options but they still boast 13 internationals including England full-back Tim Stimpson and recalled wing Tony Underwood , plus All Black forwards Ian Jones and Norm Hewitt . O O O O S-ORG O O O O O O O O O O S-LOC O B-PER E-PER O O O B-PER E-PER O O B-ORG E-ORG O B-PER E-PER O B-PER E-PER O
Teams : O O
Barbarians - 15 - Tim Stimpson ( England ) ; 14 - Nigel Walker ( Wales ) , 13 - Allan Bateman ( Wales ) , 12 - Gregor Townsend ( Scotland ) , 11 - Tony Underwood ( England ) ; 10 - Rob Andrew ( England ) , 9 - Rob Howley ( Wales ) ; 8 - Scott Quinnell ( Wales ) , 7 - Neil Back ( England ) , 6 - Dale McIntosh ( Pontypridd ) , 5 - Ian Jones ( New Zealand ) , 4 - Craig Quinnell ( Wales ) , 3 - Darren Garforth ( Leicester ) , 2 - Norm Hewitt ( New Zealand ) , 1 - Nick Popplewell ( Ireland ) . S-ORG O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O B-LOC E-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O B-LOC E-LOC O O O O B-PER E-PER O S-LOC O O
Australia - 15 - Matthew Burke ; 14 - Joe Roff , 13 - Daniel Herbert , 12 - Tim Horan ( captain ) , 11 - David Campese ; 10 - Pat Howard , 9 - Sam Payne ; 8 - Michael Brial , 7 - David Wilson , 6 - Owen Finegan , 5 - David Giffin , 4 - Tim Gavin , 3 - Andrew Blades , 2 - Marco Caputo , 1 - Dan Crowley . S-LOC O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O
-DOCSTART- O
GOLF - ZIMBABWE OPEN SECOND ROUND SCORES . O O B-MISC E-MISC O O O O
HARARE 1996-12-06 S-LOC O
Leading second round scores in the Zimbabwe Open at the par-72 Chapman Golf Club on Friday ( South African unless stated ) : 132 Des Terblanche 65 67 133 Mark McNulty ( Zimbabwe ) 72 61 134 Steve van Vuuren 65 69 136 Nick Price ( Zimbabwe ) 68 68 , Justin Hobday 71 65 , O O O O O O B-MISC E-MISC O O O B-LOC I-LOC E-LOC O O O B-MISC E-MISC O O O O O B-PER E-PER O O O B-PER E-PER O S-LOC O O O O B-PER I-PER E-PER O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O O O
Andrew Pitts ( U.S. ) 69 67 138 Mark Cayeux ( Zimbabwe ) 69 69 , Mark Murless 71 67 139 Hennie Swart 75 64 , Andrew Park 72 67 140 Schalk van der Merwe ( Namibia ) 67 73 , Desvonde B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O O O B-PER E-PER O O O B-PER E-PER O O O B-PER I-PER I-PER E-PER O S-LOC O O O O S-PER
Botes 72 68 , Greg Reid 72 68 , Clinton Whitelaw 70 S-PER O O O B-PER E-PER O O O B-PER E-PER O
70 , Brett Liddle 75 65 , Hugh Baiocchi 73 67 141 Adilson da Silva ( Brazil ) 72 69 , Sammy Daniels 73 O O B-PER E-PER O O O B-PER E-PER O O O B-PER I-PER E-PER O S-LOC O O O O B-PER E-PER O
68 , Trevor Dodds ( Namibia ) 72 69 142 Don Robertson ( U.S. ) 73 69 , Dion Fourie 69 73 , O O B-PER E-PER O S-LOC O O O O B-PER E-PER O S-LOC O O O O B-PER E-PER O O O
Steve Waltman 72 70 , Ian Dougan 73 69 B-PER E-PER O O O B-PER E-PER O O
-DOCSTART- O
SOCCER - UNCAPPED PLAYERS CALLED TO FACE MACEDONIA . O O O O O O O S-LOC O
BUCHAREST 1996-12-06 S-LOC O
Romania trainer Anghel Iordanescu called up three uncapped players on Friday in his squad to face Macedonia next week in a World Cup qualifier . S-LOC O B-PER E-PER O O O O O O O O O O O O S-LOC O O O O B-MISC E-MISC O O
Midfielder Valentin Stefan and striker Viorel Ion of Otelul Galati and defender Liviu Ciobotariu of National Bucharest are the newcomers for the European group eight clash in Macedonia on December 14 . O B-PER E-PER O O B-PER E-PER O B-ORG E-ORG O O B-PER E-PER O B-ORG E-ORG O O O O O S-MISC O O O O S-LOC O O O O
Iordanescu said he had picked them because of their good performances in the domestic championship in which National Bucharest are top and Otelul Galati third . " S-PER O O O O O O O O O O O O O O O O B-ORG E-ORG O O O B-ORG E-ORG O O O
I think it 's fair to give them a chance , " he told reporters . O O O O O O O O O O O O O O O O
League title-holders Steaua Bucharest , who finished bottom of their Champions ' League group in the European Cup , have only two players in the squad . O O B-ORG E-ORG O O O O O O B-MISC I-MISC E-MISC O O O B-MISC E-MISC O O O O O O O O O
Attacking midfielder Adrian Ilie , who recently moved from Steaua to Turkish club Galatasaray , is ruled out after two yellow-card offences . O O B-PER E-PER O O O O O S-ORG O S-MISC O S-ORG O O O O O O O O O
Squad : O O
Goalkeepers - Bogdan Stelea , Florin Prunea . O O B-PER E-PER O B-PER E-PER O
Defenders - Dan Petrescu , Daniel Prodan , Anton Dobos , Cornel Papura , Liviu Ciobotariu , Tibor Selymess , Iulian Filipescu . O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Midfielders - Gheorghe Hagi , Gheorghe Popescu , Constantin Galca , Valentin Stefan , Basarab Panduru , Dorinel Munteanu , Ovidiu Stinga . O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Forwards - Ioan Vladoiu , Gheorghe Craioveanu , Ionel Danciulescu , Viorel Ion . O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
REUTER S-ORG
-DOCSTART- O
SOCCER - BRAZILIAN CHAMPIONSHIP RESULTS . O O S-MISC O O O
RIO DE JANEIRO 1996-12-05 B-LOC I-LOC E-LOC O
Results of Brazilian O O S-MISC
soccer championship semifinal , first leg matches on Thursday . O O O O O O O O O O
Goias 1 Gremio 3 S-ORG O S-ORG O
Portuguesa 1 Atletico Mineiro 0 S-ORG O B-ORG E-ORG O
-DOCSTART- O
CRICKET - LARA ENDURES ANOTHER MISERABLE DAY . O O S-PER O O O O O
Robert Galvin B-PER E-PER
MELBOURNE 1996-12-06 S-LOC O
Australia gave Brian Lara another reason to be miserable when they beat West Indies by five wickets in the opening World Series limited overs match on Friday . S-LOC O B-PER E-PER O O O O O O O O B-LOC E-LOC O O O O O O B-MISC E-MISC O O O O O O
Lara , disciplined for misconduct on Wednesday , was dismissed for five to extend a disappointing run of form on tour . S-PER O O O O O O O O O O O O O O O O O O O O O
Australia , who hold a 2-0 lead in the five-match test series , overhauled West Indies ' total of 172 all out with eight balls to spare to end a run of six successive one-day defeats . S-LOC O O O O O O O O O O O O O B-LOC E-LOC O O O O O O O O O O O O O O O O O O O O O
All-rounder Greg Blewett steered his side to a comfortable victory with an unbeaten 57 in 90 balls to the delight of the 42,442 crowd . O B-PER E-PER O O O O O O O O O O O O O O O O O O O O O O
Man-of-the match Blewett came to the wicket with the total on 70 for two and hit three fours during an untroubled innings lasting 129 minutes . O O S-PER O O O O O O O O O O O O O O O O O O O O O O O
His crucial fifth-wicket partnership with fellow all-rounder Stuart Law , who scored 21 , added 71 off 85 balls . O O O O O O O B-PER E-PER O O O O O O O O O O O
Lara looked out of touch during his brief stay at the crease before chipping a simple catch to Shane Warne at mid-wicket . S-PER O O O O O O O O O O O O O O O O O B-PER E-PER O O O
West Indies tour manager Clive Lloyd has apologised for Lara 's behaviour on Tuesday . B-LOC E-LOC O O B-PER E-PER O O O S-PER O O O O O
He ( Lara ) had told Australia coach Geoff Marsh that wicketkeeper Ian Healy was unwelcome in the visitors ' dressing room . O O S-PER O O O S-LOC O B-PER E-PER O O B-PER E-PER O O O O O O O O O
The Melbourne crowd were clearly angered by the incident , loudly jeering the West Indies vice-captain as he walked to the middle . O S-LOC O O O O O O O O O O O B-LOC E-LOC O O O O O O O O
It was left to fellow left-hander Shivnarine Chanderpaul to hold the innings together with a gritty 54 despite the handicap of an injured groin . O O O O O O B-PER E-PER O O O O O O O O O O O O O O O O O
Chanderpaul was forced to rely on a runner for most of his innings after hurting himself as he scurried back to his crease to avoid being run out . S-PER O O O O O O O O O O O O O O O O O O O O O O O O O O O O
Pakistan , who arrive in Australia later this month , are the other team competing in the World Series tournament . S-LOC O O O O S-LOC O O O O O O O O O O O B-MISC E-MISC O O
-DOCSTART- O
CRICKET - AUSTRALIA V WEST INDIES WORLD SERIES SCOREBOARD . O O S-LOC O B-LOC E-LOC B-MISC E-MISC O O
MELBOURNE 1996-12-06 S-LOC O
Scoreboard in the World Series O O O B-MISC E-MISC
limited overs match between Australia and West Indies on Friday : O O O O S-LOC O B-LOC E-LOC O O O
West Indies B-LOC E-LOC
S. Campbell c Healy b Gillespie 31 B-PER E-PER O S-PER O S-PER O
R. Samuels c M. Waugh b Gillespie 7 B-PER E-PER O B-PER E-PER O S-PER O
B. Lara c Warne b Moody 5 B-PER E-PER O S-PER O S-PER O
S. Chanderpaul c Healy b Blewett 54 B-PER E-PER O S-PER O S-PER O
C. Hooper run out 7 B-PER E-PER O O O
J. Adams lbw b Moody 5 B-PER E-PER O O S-PER O
J. Murray c Blewett b Warne 24 B-PER E-PER O S-PER O S-PER O
N. McLean c and b M. Waugh 7 B-PER E-PER O O O B-PER E-PER O
K. Benjamin b Warne 8 B-PER E-PER O S-PER O
C. Ambrose run out 2 B-PER E-PER O O O
C. Walsh not out 8 B-PER E-PER O O O
Extras ( lb-10 w-1 nb-3 ) 14 O O O O O O O
Total ( 49.2 overs ) 172 O O O O O O
Fall of wickets : 1-11 2-38 3-64 4-73 5-81 6-120 7-135 8-150 O O O O O O O O O O O O
9-153 . O O
Bowling : Reiffel 10-2-26-0 ( nb-3 ) , Gillespie 10-0-39-2 , O O S-PER O O O O O S-PER O O
Moody 10-1-25-2 , Blewett 6.2-0-27-1 , Warne 10-0-34-2 ( w-1 ) , S-PER O O S-PER O O S-PER O O O O O
M. Waugh 3-0-11-1 . B-PER E-PER O O
Australia S-LOC
M. Taylor b McLean 29 B-PER E-PER O S-PER O
M. Waugh c Murray b Benjamin 27 B-PER E-PER O S-PER O S-PER O
R. Ponting lbw McLean 5 B-PER E-PER O S-PER O
G. Blewett not out 57 B-PER E-PER O O O
M. Bevan st Murray b Hooper 3 B-PER E-PER O S-PER O S-PER O
S. Law b Hooper 21 B-PER E-PER O S-PER O
T. Moody not out 3 B-PER E-PER O O O
Extras ( lb-17 nb-8 w-3 ) 28 O O O O O O O
Total ( for five wickets , 48.4 overs ) 173 O O O O O O O O O O
Fall of wickets : 1-59 2-70 3-78 4-90 5-160 . O O O O O O O O O O
Did not bat : I. Healy , P. Reiffel , S. Warne , J. Gillespie . O O O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O
Bowling : Ambrose 10-3-19-0 ( 2nb 1w ) , Walsh 9-0-34-0 ( 4nb ) , O O S-PER O O O O O O S-PER O O O O O
Benjamin 9.4-0-43-1 ( 1nb 1w ) , Hooper 10-0-27-2 ( 1nb ) , McLean S-PER O O O O O O S-PER O O O O O S-PER
10-1-33-2 ( 1w ) . O O O O O
Result : Australia won by five wickets . O O S-LOC O O O O O
-DOCSTART- O
CRICKET - AUSTRALIA BEAT WEST INDIES BY FIVE WICKETS . O O S-LOC O B-LOC E-LOC O O O O
MELBOURNE 1996-12-06 S-LOC O
Australia beat West Indies by five wickets in a World Series limited overs match at the Melbourne Cricket Ground on Friday . S-LOC O B-LOC E-LOC O O O O O B-MISC E-MISC O O O O O B-LOC I-LOC E-LOC O O O
Scores : West Indies 172 all out in 49.2 overs ( Shivnarine Chanderpaul 54 ) ; Australia 173-5 in 48.4 overs ( Greg Blewett 57 not out ) . O O B-LOC E-LOC O O O O O O O B-PER E-PER O O O S-LOC O O O O O B-PER E-PER O O O O O
-DOCSTART- O
CRICKET - WEST INDIES 172 ALL OUT IN 49.2 OVERS V AUSTRALIA . O O B-LOC E-LOC O O O O O O O S-LOC O
MELBOURNE 1996-12-06 S-LOC O
West Indies were all out for 172 off 49.2 overs in the World Series limited overs match against Australia on Friday . B-LOC E-LOC O O O O O O O O O O B-MISC E-MISC O O O O S-LOC O O O
-DOCSTART- O
CRICKET - SHEFFIELD SHIELD SCORE . O O B-MISC E-MISC O O
HOBART , Australia 1996-12-06 S-LOC O S-LOC O
Score on the first day of the four-day Sheffield Shield match between Tasmania and Victoria at Bellerive Oval on Friday : O O O O O O O O B-MISC E-MISC O O S-LOC O S-LOC O B-LOC E-LOC O O O
Tasmania 352 for three ( David Boon 106 not out , Shaun Young 86 not out , Michael DiVenuto 119 ) v Victoria . S-LOC O O O O B-PER E-PER O O O O B-PER E-PER O O O O B-PER E-PER O O O S-ORG O
-DOCSTART- O
CRICKET - LARA SUFFERS MORE AUSTRALIAN TOUR MISERY . O O S-PER O O O O O O
MELBOURNE 1996-12-06 S-LOC O
West Indies batsman Brian Lara suffered another blow to his Australian tour , after already being disciplined for misconduct , when he was dismissed cheaply in the first limited overs match against Australia on Friday . B-LOC E-LOC O B-PER E-PER O O O O O S-MISC O O O O O O O O O O O O O O O O O O O O O S-LOC O O O
Lara , who earned a stern rebuke from his own tour management after an angry outburst against Australia wicketkeeper Ian Healy , scored five to prolong a run of poor form with the bat . S-PER O O O O O O O O O O O O O O O O S-LOC O B-PER E-PER O O O O O O O O O O O O O O
The West Indies vice-captain struggled for timing during his 36-minute stay at the crease before chipping a ball from medium pacer Tom Moody straight to Shane Warne at mid-wicket . O B-LOC E-LOC O O O O O O O O O O O O O O O O O O B-PER E-PER O O B-PER E-PER O O O
West Indies were 53 for two in 15 overs when rain stopped play at the Melbourne Cricket Ground after captain Courtney Walsh won the toss and elected to bat . B-LOC E-LOC O O O O O O O O O O O O O B-LOC I-LOC E-LOC O O B-PER E-PER O O O O O O O O
Lara 's outburst three days ago has clearly turned some of the Australian public against him . S-PER O O O O O O O O O O O S-MISC O O O O
As he walked to the wicket he was greeted by loud jeers from sections of the crowd . O O O O O O O O O O O O O O O O O O
On several occasions during his innings , the crowd joined together in a series of obscene chants against him . O O O O O O O O O O O O O O O O O O O O
Tour manager Clive Lloyd on Wednesday apologised for Lara 's behaviour in confronting Australia coach Geoff Marsh in the opposition dressing room to protest against his dismissal in the second test on Tuesday . O O B-PER E-PER O O O O S-PER O O O O S-LOC O B-PER E-PER O O O O O O O O O O O O O O O O O
Lloyd did not say what form the discipline would take . S-PER O O O O O O O O O O
Lara , who holds the record for the highest score in test and first-class cricket , was unhappy about Healy 's role in the incident and questioned whether the ball had carried to the Australia keeper . S-PER O O O O O O O O O O O O O O O O O O S-PER O O O O O O O O O O O O O O S-LOC O O
Australia went on to win the match at the Sydney Cricket Ground by 124 runs to take a two-nil lead in the five-test series after Lara failed in both innings . S-LOC O O O O O O O O B-LOC I-LOC E-LOC O O O O O O O O O O O O O S-PER O O O O O
Lara has yet to score a century since West Indies arrived in Australia five weeks ago . S-PER O O O O O O O B-LOC E-LOC O O S-LOC O O O O
Both West Indies and Australia team management have played down the incident , stressing that relations between the two sides have not been adversely affected . O B-LOC E-LOC O S-LOC O O O O O O O O O O O O O O O O O O O O O
Pakistan , who arrive next week , are the third team in the triangular World Series tournament . S-LOC O O O O O O O O O O O O O B-MISC E-MISC O O
-DOCSTART- O
CRICKET - WEST INDIES TO BAT AFTER WINNING THE TOSS . O O B-LOC E-LOC O O O O O O O
MELBOURNE 1996-12-06 S-LOC O
West Indies captain Courtney Walsh elected to bat after winning the toss in the first match in the World Series limited overs competition against Australia at the Melbourne Cricket Ground on Friday . B-LOC E-LOC O B-PER E-PER O O O O O O O O O O O O O B-MISC E-MISC O O O O S-LOC O O S-LOC O O O O O
Teams : O O
Australia - Mark Taylor ( captain ) , Mark Waugh , Ricky Ponting , Greg Blewett , Michael Bevan , Stuart Law , Tom Moody , Ian Healy , Paul Reiffel , Shane Warne , Jason Gillespie , Glenn McGrath 12th man . S-LOC O B-PER E-PER O O O O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O O O
West Indies - Sherwin Campbell , Robert Samuels , Brian Lara , Shivnarine Chanderpaul , Carl Hooper , Jimmy Adams , Junior Murray , Nixon McLean , Kenneth Benjamin , Curtly Ambrose , Courtney Walsh ( captain ) , Roland Holder 12th man . B-LOC E-LOC O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O B-PER E-PER O O O O B-PER E-PER O O O
-DOCSTART- O
BADMINTON - WORLD GRAND PRIX RESULTS . O O B-MISC I-MISC E-MISC O O
BALI 1996-12-06 S-LOC O
Results in last of the group matches at the World Grand Prix badminton finals on Friday : O O O O O O O O O B-MISC I-MISC E-MISC O O O O O
Men 's singles O O O
Group B O O
Chen Gang ( China ) beat Martin Londgaard Hansen ( Denmark ) 15-12 15-6 B-PER E-PER O S-LOC O O B-PER I-PER E-PER O S-LOC O O O
Dong Jiong ( China ) beat Thomas Stuer-Lauridsen ( Denmark ) 15-10 15-6 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Indra Wijaya ( Indonesia ) beat Ong Ewe Hock ( Malaysia ) 5-15 15-11 15-11 B-PER E-PER O S-LOC O O B-PER I-PER E-PER O S-LOC O O O O
Group C O O
Sun Jun ( China ) beat Rashid Sidek ( Malaysia ) 15-12 17-14 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Hermawan Susanto ( Indonesia ) beat Soren B. Nielsen ( Denmark ) 15-8 15-2 B-PER E-PER O S-LOC O O B-PER I-PER E-PER O S-LOC O O O
Group D O O
Allan Budi Kuksuma ( Indonesia ) beat Poul-Erik Hoyer-Larsen ( Denmark ) 15-7 15-4 B-PER I-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Budi Santoso ( Indonesia ) beat Hu Zhilan ( China ) 15-4 15-5 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Semifinals ( on Saturday ) : Fung Permadi ( Taiwan ) v Indra O O O O O O B-PER E-PER O S-LOC O O S-PER
Wijaya ( Indonesia ) ; Sun Jun ( China ) v Allan Budi Kusuma S-PER O S-LOC O O B-PER E-PER O S-LOC O O B-PER I-PER E-PER
( Indonesia ) O S-LOC O
Women 's singles O O O
Group A O O
Gong Zhichao ( China ) beat Mia Audina ( Indonesia ) 11-2 12-10 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Group B O O
Ye Zhaoying ( China ) beat Meiluawati ( Indonesia ) 11-6 12-10 B-PER E-PER O S-LOC O O S-PER O S-LOC O O O
Group C O O
Camilla Martin ( Denmark ) beat Wang Chen ( China ) 11-0 12-10 B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O
Group D O O
Susi Susanti ( Indonesia ) beat Han Jingna ( China ) 11-5 11-4 . B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O O O
Semifinals ( on Saturday ) : Susi Susanti ( Indonesia ) v Camilla Martin ( Denmark ) ; Ye Zhaoying ( China ) v Gong Zichao ( China ) . O O O O O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O B-PER E-PER O S-LOC O O
-DOCSTART- O
SOCCER - ARAB CONTRACTORS WIN AFRICAN CUP WINNERS ' CUP . O O S-MISC O O B-MISC I-MISC I-MISC I-MISC E-MISC O
CAIRO 1996-12-06 S-LOC O
Result of the second leg of the African Cup Winners ' Cup final at the National stadium on Friday : Arab Contractors ( Egypt ) 4 Sodigraf ( Zaire ) 0 ( halftime 2-0 ) O O O O O O O B-MISC I-MISC I-MISC I-MISC E-MISC O O O B-LOC E-LOC O O O B-ORG E-ORG O S-LOC O O S-ORG O S-LOC O O O O O O
Scorers : O O
Aly Ashour 7 , 56 penalty , Mohamed Ouda 24 , 73 B-PER E-PER O O O O O B-PER E-PER O O O
Contractors won 4-0 on aggregate . O O O O O O
-DOCSTART- O
NHL ICE HOCKEY - STANDINGS AFTER THURSDAY 'S GAMES . S-ORG O O O O O O O O O
NEW YORK 1996-12-06 B-LOC E-LOC O
Standings of National Hockey O O B-ORG E-ORG
League teams after games played on Thursday ( tabulate under S-ORG O O O O O O O O O
won , lost , tied , goals for , goals against , points ) : O O O O O O O O O O O O O O O
EASTERN CONFERENCE O O
NORTHEAST DIVISION O O
W L T GF GA PTS O O O O O O
HARTFORD 12 7 6 77 76 30 S-ORG O O O O O O
BUFFALO 13 12 1 77 76 27 S-ORG O O O O O O
BOSTON 10 11 4 74 84 24 S-ORG O O O O O O
MONTREAL 10 14 4 96 103 24 S-ORG O O O O O O
PITTSBURGH 9 13 3 81 91 21 S-ORG O O O O O O
OTTAWA 7 11 6 62 72 20 S-ORG O O O O O O
ATLANTIC DIVISION S-LOC O
W L T GF GA PTS O O O O O O
FLORIDA 17 4 6 83 53 40 S-ORG O O O O O O
PHILADELPHIA 14 12 2 75 75 30 S-ORG O O O O O O
NEW JERSEY 14 10 1 61 61 29 B-ORG E-ORG O O O O O O
WASHINGTON 13 12 1 69 66 27 S-ORG O O O O O O
NY RANGERS 10 13 5 91 81 25 B-ORG E-ORG O O O O O O
NY ISLANDERS 7 11 8 65 72 22 B-ORG E-ORG O O O O O O
TAMPA BAY 8 15 2 69 81 18 B-ORG E-ORG O O O O O O
WESTERN CONFERENCE O O
CENTRAL DIVISION B-MISC E-MISC
W L T GF GA PTS O O O O O O
DETROIT 15 9 4 81 53 34 S-ORG O O O O O O
DALLAS 16 9 1 74 60 33 S-ORG O O O O O O
CHICAGO 12 12 3 71 67 27 S-ORG O O O O O O
ST LOUIS 13 14 0 78 81 26 B-ORG E-ORG O O O O O O
TORONTO 11 15 0 76 89 22 S-ORG O O O O O O
PHOENIX 9 13 4 61 74 22 S-ORG O O O O O O
PACIFIC DIVISION S-LOC O
W L T GF GA PTS O O O O O O
COLORADO 17 6 4 97 56 38 S-ORG O O O O O O
VANCOUVER 14 11 1 84 83 29 S-ORG O O O O O O
EDMONTON 13 14 1 94 88 27 S-ORG O O O O O O
LOS ANGELES 11 13 3 72 83 25 B-ORG E-ORG O O O O O O
SAN JOSE 10 13 4 69 87 24 B-ORG E-ORG O O O O O O
CALGARY 10 16 2 65 77 22 S-ORG O O O O O O
ANAHEIM 9 14 4 73 86 22 S-ORG O O O O O O
FRIDAY , DECEMBER 6 O O O O
ANAHEIM AT BUFFALO S-ORG O S-LOC
TORONTO AT NY RANGERS S-ORG O B-ORG E-ORG
PITTSBURGH AT WASHINGTON S-ORG O S-LOC
MONTREAL AT CHICAGO S-ORG O S-LOC
PHILADELPHIA AT DALLAS S-ORG O S-LOC
ST LOUIS AT COLORADO B-ORG E-ORG O S-LOC
OTTAWA AT EDMONTON S-ORG O S-LOC
-DOCSTART- O
NHL ICE HOCKEY - THURSDAY 'S RESULTS . S-ORG O O O O O O O
[ CORRECTED 08:40 GMT ] O O O S-MISC O
NEW YORK 1996-12-06 B-LOC E-LOC O
( Corrects headline from NBA to NHL and corrects team name in second result from La Clippers to Ny Islanders . O O O O S-ORG O S-ORG O O O O O O O O B-ORG E-ORG O B-ORG E-ORG O
) O
Results of National Hockey O O B-ORG E-ORG
League games on Thursday ( home team in CAPS ) : S-ORG O O O O O O O O O O
Hartford 4 BOSTON 2 S-ORG O S-ORG O
FLORIDA 4 Ny Islanders 2 S-ORG O B-ORG E-ORG O
NEW JERSEY 2 Calgary 1 B-ORG E-ORG O S-ORG O
Phoenix 3 ST LOUIS 0 S-ORG O B-ORG E-ORG O
Tampa Bay 2 LOS ANGELES 1 B-ORG E-ORG O B-ORG E-ORG O
-DOCSTART- O
NFL AMERICAN FOOTBALL-COLTS CLOBBER EAGLES TO STAY IN PLAYOFF HUNT . S-ORG O O O S-ORG O O O O O O
INDIANAPOLIS 1996-12-06 S-LOC O
The injury-plagued Indianapolis Colts lost another quarterback on Thursday but last year 's AFC finalists rallied together to shoot down the Philadelphia Eagles 37-10 in a showdown of playoff contenders . O O B-ORG E-ORG O O O O O O O O O O O O O O O O O B-ORG E-ORG O O O O O O O O
Marshall Faulk rushed for 101 yards and two touchdowns and Jason Belser returned an interception 44 yards for a score as the Colts improved to 8-6 , the same mark as the Eagles , who lost for the fourth time in five games . B-PER E-PER O O O O O O O O B-PER E-PER O O O O O O O O O O S-ORG O O O O O O O O O S-ORG O O O O O O O O O O O
Paul Justin , starting for the sidelined Jim Harbaugh , was 14-of-23 for 144 yards and a touchdown for the the Colts , who played their last home game of the season . B-PER E-PER O O O O O B-PER E-PER O O O O O O O O O O O O S-ORG O O O O O O O O O O O
Indianapolis closes with games at Kansas City and Cincinnati . S-LOC O O O O B-LOC E-LOC O S-LOC O
The Eagles were held without a touchdown until the final five seconds . O S-ORG O O O O O O O O O O O
Philadelphia , which fell from an NFC East tie with the Dallas Cowboys and Washington Redskins , go on the road against the New York Jets and then entertain Arizona . S-LOC O O O O O O O O O O B-ORG E-ORG O B-ORG E-ORG O O O O O O O B-ORG I-ORG E-ORG O O O S-ORG O
The loss by Philadelphia allowed the idle Green Bay Packers ( 10-3 ) to clinch the first NFC playoff berth . O O O S-ORG O O O B-ORG I-ORG E-ORG O O O O O O O O O O O
The Colts won despite the absence of injured starting defensive tackle Tony Siragusa , cornerback Ray Buchanan and linebacker Quentin Coryatt . O S-ORG O O O O O O O O O B-PER E-PER O O B-PER E-PER O O B-PER E-PER O
Faulk carried 16 times , including a 13-yard TD run in the first quarter and a seven-yard score early in the final period . S-PER O O O O O O O O O O O O O O O O O O O O O O O
Justin made his second straight start for Harbaugh , who has a knee injury . S-PER O O O O O O S-PER O O O O O O O
Justin suffered a sprained right shoulder in the third quarter and did not return . S-PER O O O O O O O O O O O O O O
Third-stringer Kerwin Bell , a 1988 draft choice of the Miami Dolphins , made his NFL debut and was 5-of-5 for 75 yards , including a 20-yard scoring strike to Marvin Harrison in the third period . O B-PER E-PER O O O O O O O B-ORG E-ORG O O O S-ORG O O O O O O O O O O O O O O B-PER E-PER O O O O O
A 39-yard interference penalty against Philadelphia 's Troy Vincent set up Faulk 's first score around left end that capped an 80-yard march 5:17 into the game and the rout was on . O O O O O S-LOC O B-PER E-PER O O S-PER O O O O O O O O O O O O O O O O O O O O O
Eagles quarterback Ty Detmer was 17-of-34 for 182 yards before he was benched . S-ORG O B-PER E-PER O O O O O O O O O O
Ricky Watters , who leads the NFC in rushing , left the game after getting kneed to the helmet after gaining 33 yards on seven carries . B-PER E-PER O O O O O O O O O O O O O O O O O O O O O O O O O
-DOCSTART- O
NBA BASKETBALL - STANDINGS AFTER THURSDAY 'S GAMES . S-ORG O O O O O O O O
NEW YORK 1996-12-06 B-LOC E-LOC O
Standings of National O O S-ORG
Basketball Association teams after games played on Thursday B-ORG E-ORG O O O O O O
( tabulate under won , lost , percentage , games behind ) : O O O O O O O O O O O O O
EASTERN CONFERENCE O O
ATLANTIC DIVISION S-LOC O
W L PCT GB O O O O
MIAMI 14 4 .778 - S-ORG O O O O
NEW YORK 10 6 .625 3 B-ORG E-ORG O O O O
ORLANDO 8 6 .571 4 S-ORG O O O O
WASHINGTON 7 9 .438 6 S-ORG O O O O
PHILADELPHIA 7 10 .412 6 1/2 S-ORG O O O O O
BOSTON 4 12 .250 9 S-ORG O O O O
NEW JERSEY 3 10 .231 8 1/2 B-ORG E-ORG O O O O O
CENTRAL DIVISION O O
W L PCT GB O O O O
CHICAGO 17 1 .944 - S-ORG O O O O
DETROIT 13 3 .813 3 S-ORG O O O O
CLEVELAND 11 5 .688 5 S-ORG O O O O
ATLANTA 10 8 .556 7 S-ORG O O O O
CHARLOTTE 8 8 .500 8 S-ORG O O O O
MILWAUKEE 8 8 .500 8 S-ORG O O O O
INDIANA 7 8 .467 8 1/2 S-ORG O O O O O
TORONTO 6 11 .353 10 1/2 S-ORG O O O O O
WESTERN CONFERENCE O O
MIDWEST DIVISION O O
W L PCT GB O O O O
HOUSTON 16 2 .889 - S-ORG O O O O
UTAH 14 2 .875 1 S-ORG O O O O
MINNESOTA 7 10 .412 8 1/2 S-ORG O O O O O
DALLAS 6 11 .353 9 1/2 S-ORG O O O O O
DENVER 5 14 .263 11 1/2 S-ORG O O O O O
SAN ANTONIO 3 13 .188 12 B-ORG E-ORG O O O O
VANCOUVER 2 16 .111 14 S-ORG O O O O
PACIFIC DIVISION S-LOC O
W L PCT GB O O O O
SEATTLE 15 5 .750 - S-ORG O O O O
LA LAKERS 13 7 .650 2 B-ORG E-ORG O O O O
PORTLAND 11 8 .579 3 1/2 S-ORG O O O O O
LA CLIPPERS 7 11 .389 7 B-ORG E-ORG O O O O
GOLDEN STATE 6 12 .333 8 B-ORG E-ORG O O O O
Не удается отобразить этот файл, потому что он содержит неожиданный символ в строке 33 и столбце 46.

Разница между файлами не показана из-за своего большого размера Загрузить разницу

71
losses/CRFLoss.py Normal file
Просмотреть файл

@ -0,0 +1,71 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import torch
import torch.nn as nn
import torch.autograd as autograd
class CRFLoss(nn.Module):
"""CRFLoss
use for crf output layer for sequence tagging task.
"""
def __init__(self):
super(CRFLoss, self).__init__()
def _score_sentence(self, scores, mask, tags, transitions, crf_layer_conf):
"""
input:
scores: variable (seq_len, batch, tag_size, tag_size)
mask: (batch, seq_len)
tags: tensor (batch, seq_len)
output:
score: sum of score for gold sequences within whole batch
"""
# Gives the score of a provided tag sequence
batch_size = scores.size(1)
seq_len = scores.size(0)
tag_size = scores.size(2)
# convert tag value into a new format, recorded label bigram information to index
new_tags = autograd.Variable(torch.LongTensor(batch_size, seq_len))
if crf_layer_conf.use_gpu:
new_tags = new_tags.cuda()
for idx in range(seq_len):
if idx == 0:
# start -> first score
new_tags[:, 0] = (tag_size-2)*tag_size + tags[:, 0]
else:
new_tags[:, idx] = tags[:, idx-1]*tag_size + tags[:, idx]
# transition for label to STOP_TAG
end_transition = transitions[:, crf_layer_conf.target_dict[crf_layer_conf.STOP_TAG]].contiguous().view(1, tag_size).expand(batch_size, tag_size)
# length for batch, last word position = length - 1
length_mask = torch.sum(mask.long(), dim=1).view(batch_size, 1).long()
# index the label id of last word
end_ids = torch.gather(tags, 1, length_mask - 1)
# index the transition score for end_id to STOP_TAG
end_energy = torch.gather(end_transition, 1, end_ids)
# convert tag as (seq_len, batch_size, 1)
new_tags = new_tags.transpose(1, 0).contiguous().view(seq_len, batch_size, 1)
# need convert tags id to search from positions of scores
tg_energy = torch.gather(scores.view(seq_len, batch_size, -1), 2, new_tags).view(seq_len, batch_size) # seq_len * batch_size
# mask transpose to (seq_len, batch_size)
tg_energy = tg_energy.masked_select(mask.transpose(1, 0))
# add all score together
gold_score = tg_energy.sum() + end_energy.sum()
return gold_score
def forward(self, forward_score, scores, masks, tags, transitions, crf_layer_conf):
"""
:param forward_score: Tensor scale
:param scores: Tensor [seq_len, batch_size, target_size, target_size]
:param masks: Tensor [batch_size, seq_len]
:param tags: Tensor [batch_size, seq_len]
:return: goal_score - forward_score
"""
gold_score = self._score_sentence(scores, masks, tags, transitions, crf_layer_conf)
return forward_score - gold_score

Просмотреть файл

@ -8,7 +8,7 @@ import logging
sys.path.append('../')
from settings import LossOperationType
from torch.nn import CrossEntropyLoss, L1Loss, MSELoss, NLLLoss, PoissonNLLLoss, NLLLoss2d, KLDivLoss, BCELoss, BCEWithLogitsLoss, MarginRankingLoss, HingeEmbeddingLoss, MultiLabelMarginLoss, SmoothL1Loss, SoftMarginLoss, MultiLabelSoftMarginLoss, CosineEmbeddingLoss, MultiMarginLoss, TripletMarginLoss
from .CRFLoss import CRFLoss
class Loss(nn.Module):
'''

Просмотреть файл

@ -1,5 +1,6 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
from .FocalLoss import FocalLoss
from .CRFLoss import CRFLoss
from .Loss import Loss
from torch.nn import CrossEntropyLoss, L1Loss, MSELoss, NLLLoss, PoissonNLLLoss, NLLLoss2d, KLDivLoss, BCELoss, BCEWithLogitsLoss, MarginRankingLoss, HingeEmbeddingLoss, MultiLabelMarginLoss, SmoothL1Loss, SoftMarginLoss, MultiLabelSoftMarginLoss, CosineEmbeddingLoss, MultiMarginLoss, TripletMarginLoss

Просмотреть файл

@ -4,6 +4,7 @@
from sklearn import metrics
from sklearn.metrics import mean_squared_error
from .conlleval import countChunks, evaluate, to_conll_format
from .slot_tagging_metrics import get_ner_BIOES, get_ner_BIO
from settings import TaggingSchemes
import numpy as np
import re
@ -138,24 +139,72 @@ class Evaluator(object):
def accuracy(self, y_true, y_pred):
return metrics.accuracy_score(y_true, y_pred)
def seq_tag_f1(self, y_true, y_pred):
""" For sequence tagging task, calculate F1-score(e.g. CONLL 2000)
def seq_tag_f1(self, y_ture, y_pred):
'''
Args:
y_true:
y_pred:
Returns:
"""
:param y_ture:
:param y_pred:
:return:
'''
assert self.__tagging_scheme is not None, "Please define tagging scheme!"
if TaggingSchemes[self.__tagging_scheme] == TaggingSchemes.BIO:
result_conll_format = to_conll_format(y_true, y_pred)
correctChunk, foundGuessed, foundCorrect, correctTags, tokenCounter = countChunks(result_conll_format)
overall_precision, overall_recall, overall_FB1 = evaluate(correctChunk, foundGuessed, foundCorrect, correctTags, tokenCounter)
sent_num = len(y_pred)
golden_full = []
predict_full = []
right_full = []
for idx in range(0, sent_num):
golden_list = y_ture[idx]
predict_list = y_pred[idx]
if self.__tagging_scheme == "BMES" or self.__tagging_scheme == "BIOES":
gold_matrix = get_ner_BIOES(golden_list)
pred_matrix = get_ner_BIOES(predict_list)
elif self.__tagging_scheme == "BIO":
gold_matrix = get_ner_BIO(golden_list)
pred_matrix = get_ner_BIO(predict_list)
else:
raise Exception("TO DO: SUPPORT MORE TAGGING SCHEMES")
return overall_FB1
# raise Exception("DETECT UNKNOWN TAGGING SCHEMES! YOU CAN USE OUR SCRIPT TO CONVERT TAG SCHEME!")
raise Exception("DETECT UNKNOWN TAGGING SCHEMES!")
right_ner = list(set(gold_matrix).intersection(set(pred_matrix)))
golden_full += gold_matrix
predict_full += pred_matrix
right_full += right_ner
right_num = len(right_full)
golden_num = len(golden_full)
predict_num = len(predict_full)
if predict_num == 0:
precision = -1
else:
precision = (right_num + 0.0) / predict_num
if golden_num == 0:
recall = -1
else:
recall = (right_num + 0.0) / golden_num
if (precision == -1) or (recall == -1) or (precision + recall) <= 0.:
f_measure = -1
else:
f_measure = 2 * precision * recall / (precision + recall)
return f_measure
def seq_tag_accuracy(self, y_ture, y_pred):
'''
:param y_ture:
:param y_pred:
:return:
'''
sent_num = len(y_pred)
right_tag = 0
all_tag = 0
for idx in range(0, sent_num):
golden_list = y_ture[idx]
predict_list = y_pred[idx]
for idy in range(len(golden_list)):
if golden_list[idy] == predict_list[idy]:
right_tag += 1
all_tag += len(golden_list)
accuracy = (right_tag + 0.0) / all_tag
return accuracy
def macro_f1(self, y_true, y_pred):
""" For classification task, calculate f1-score for each label, and find their unweighted mean. This does not take label imbalance into account.

Просмотреть файл

@ -0,0 +1,97 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
def get_ner_BIOES(label_list):
list_len = len(label_list)
begin_label = 'B-'
end_label = 'E-'
single_label = 'S-'
whole_tag = ''
index_tag = ''
tag_list = []
stand_matrix = []
for i in range(0, list_len):
current_label = label_list[i].upper()
if begin_label in current_label:
if index_tag != '':
tag_list.append(whole_tag + ',' + str(i-1))
whole_tag = current_label.replace(begin_label, "", 1) + '[' + str(i)
index_tag = current_label.replace(begin_label, "", 1)
elif single_label in current_label:
if index_tag != '':
tag_list.append(whole_tag + ',' + str(i-1))
whole_tag = current_label.replace(single_label, "", 1) + '[' +str(i)
tag_list.append(whole_tag)
whole_tag = ""
index_tag = ""
elif end_label in current_label:
if index_tag != '':
tag_list.append(whole_tag + ',' + str(i))
whole_tag = ''
index_tag = ''
else:
continue
if (whole_tag != '')&(index_tag != ''):
tag_list.append(whole_tag)
tag_list_len = len(tag_list)
for i in range(0, tag_list_len):
if len(tag_list[i]) > 0:
tag_list[i] = tag_list[i]+ ']'
insert_list = reverse_style(tag_list[i])
stand_matrix.append(insert_list)
return stand_matrix
def get_ner_BIO(label_list):
list_len = len(label_list)
begin_label = 'B-'
inside_label = 'I-'
whole_tag = ''
index_tag = ''
tag_list = []
stand_matrix = []
for i in range(0, list_len):
current_label = label_list[i].upper()
if begin_label in current_label:
if index_tag == '':
whole_tag = current_label.replace(begin_label, "", 1) + '[' + str(i)
index_tag = current_label.replace(begin_label, "", 1)
else:
tag_list.append(whole_tag + ',' + str(i-1))
whole_tag = current_label.replace(begin_label, "", 1) + '[' + str(i)
index_tag = current_label.replace(begin_label, "", 1)
elif inside_label in current_label:
if current_label.replace(inside_label, "", 1) == index_tag:
whole_tag = whole_tag
else:
if (whole_tag != '')&(index_tag != ''):
tag_list.append(whole_tag + ',' + str(i-1))
whole_tag = ''
index_tag = ''
else:
if (whole_tag != '')&(index_tag != ''):
tag_list.append(whole_tag + ',' + str(i-1))
whole_tag = ''
index_tag = ''
if (whole_tag != '')&(index_tag != ''):
tag_list.append(whole_tag)
tag_list_len = len(tag_list)
for i in range(0, tag_list_len):
if len(tag_list[i]) > 0:
tag_list[i] = tag_list[i]+ ']'
insert_list = reverse_style(tag_list[i])
stand_matrix.append(insert_list)
return stand_matrix
def reverse_style(input_string):
target_position = input_string.index('[')
input_len = len(input_string)
output_string = input_string[target_position:input_len] + input_string[0:target_position]
return output_string

Просмотреть файл

@ -49,7 +49,8 @@
"training_params": {
"vocabulary": {
"min_word_frequency": 1,
"max_vocabulary": 100000
"max_vocabulary": 100000,
"max_building_lines": 1000000
},
"optimizer": {
"name": "Adam",
@ -57,6 +58,7 @@
"lr": 0.001
}
},
"chunk_size": 1000000,
"lr_decay": 0.95,
"minimum_lr": 0.0001,
"epoch_start_lr_decay": 1,
@ -65,7 +67,7 @@
"batch_size": 30,
"batch_num_to_show_results": 10,
"max_epoch": 3,
"valid_times_per_epoch": 1,
"steps_per_validation": 10,
"text_preprocessing": ["DBC2SBC"],
"max_lengths":{
"question": 30,
@ -75,6 +77,7 @@
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": false,
"conf": {
"word": {
"cols": ["question_text", "answer_text"],
@ -89,11 +92,11 @@
"cols": ["question_char", "answer_char"],
"type": "CNNCharEmbedding",
"dropout": 0.2,
"dim": 30,
"embedding_matrix_dim": 8,
"stride":1,
"window_size": 5,
"activation": null
"dim": [30, 20, 100],
"embedding_matrix_dim": 50,
"stride":[1, 2, 3],
"window_size": [3,3,5],
"activation": "ReLU"
}
}
},

Просмотреть файл

@ -53,7 +53,7 @@
"batch_size": 256,
"batch_num_to_show_results": 10,
"max_epoch": 30,
"valid_times_per_epoch": 10,
"steps_per_validation": 10,
"fixed_lengths":{
"query": 30
}

Просмотреть файл

@ -0,0 +1,241 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for question answer matching task, and it achieved auc: 0.7508 in WikiQACorpus test set",
"language": "English",
"inputs": {
"use_cache": true,
"dataset_type": "classification",
"data_paths": {
"train_data_path": "./dataset/WikiQACorpus/WikiQA-train.tsv",
"valid_data_path": "./dataset/WikiQACorpus/WikiQA-dev.tsv",
"test_data_path": "./dataset/WikiQACorpus/WikiQA-test.tsv",
"pre_trained_emb": "./dataset/Glove/glove.840B.300d.txt"
},
"file_with_col_header": true,
"add_start_end_for_seq": true,
"file_header": {
"question_id": 0,
"question_text": 1,
"document_id": 2,
"document_title": 3,
"passage_id": 4,
"passage_text": 5,
"label": 6
},
"model_inputs": {
"question": ["question_text"],
"passage": ["passage_text"]
},
"target": ["label"]
},
"outputs":{
"save_base_dir": "./models/wikiqa_arci/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.wikiqa_arci/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"optimizer": {
"name": "Adam",
"params": {
"lr": 0.001
}
},
"fixed_lengths": {
"question": 200,
"passage": 200
},
"lr_decay": 0.90,
"minimum_lr": 0.00005,
"epoch_start_lr_decay": 20,
"use_gpu": true,
"cpu_num_workers": 1,
"batch_size": 64,
"batch_num_to_show_results": 500,
"max_epoch": 10,
"valid_times_per_epoch": 2
},
"architecture":[
{
"layer": "Embedding",
"conf": {
"word": {
"cols": ["question_text", "passage_text"],
"dim": 300,
"fix_weight": false
}
}
},
{
"layer_id": "s1_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["question"]
},
{
"layer_id": "s2_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["passage"]
},
{
"layer_id": "s1_conv_1",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME",
"remind_lengths": false
},
"inputs": ["s1_dropout"]
},
{
"layer_id": "s1_pool_1",
"layer": "Pooling1D",
"conf": {
"stride": 1,
"window_size": 2
},
"inputs": ["s1_conv_1"]
},
{
"layer_id": "s1_conv_2",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME"
},
"inputs": ["s1_pool_1"]
},
{
"layer_id": "s1_pool_2",
"layer": "Pooling1D",
"conf": {
"stride": 1,
"window_size": 2
},
"inputs": ["s1_conv_2"]
},
{
"layer_id": "s1_flatten",
"layer": "Flatten",
"conf": {
},
"inputs": ["s1_pool_2"]
},
{
"layer_id": "s2_conv_1",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME",
"remind_lengths": false
},
"inputs": ["s2_dropout"]
},
{
"layer_id": "s2_pool_1",
"layer": "Pooling1D",
"conf": {
"stride": 1,
"window_size": 2
},
"inputs": ["s2_conv_1"]
},
{
"layer_id": "s2_conv_2",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME"
},
"inputs": ["s2_pool_1"]
},
{
"layer_id": "s2_pool_2",
"layer": "Pooling1D",
"conf": {
"stride": 1,
"window_size": 2
},
"inputs": ["s2_conv_2"]
},
{
"layer_id": "s2_flatten",
"layer": "Flatten",
"conf": {
},
"inputs": ["s2_pool_2"]
},
{
"layer_id": "comb",
"layer": "Combination",
"conf": {
"operations": ["origin"]
},
"inputs": ["s1_flatten", "s2_flatten"]
},
{
"layer_id": "comb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["comb"]
},
{
"layer_id": "mlp",
"layer": "Linear",
"conf": {
"hidden_dim": [64, 32],
"activation": "ReLU",
"batch_norm": true,
"last_hidden_activation": true
},
"inputs": ["comb_dropout"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "ReLU",
"batch_norm": true,
"last_hidden_activation": false,
"last_hidden_softmax": true
},
"inputs": ["mlp"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"weight": [0.1,0.9],
"size_average": true
},
"inputs": ["output","label"]
}
]
},
"metrics": ["auc", "accuracy"]
}

Просмотреть файл

@ -0,0 +1,212 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for question answer matching task, and it achieved auc: 0.7612 in WikiQACorpus test set",
"language": "English",
"inputs": {
"use_cache": true,
"dataset_type": "classification",
"data_paths": {
"train_data_path": "./dataset/WikiQACorpus/WikiQA-train.tsv",
"valid_data_path": "./dataset/WikiQACorpus/WikiQA-dev.tsv",
"test_data_path": "./dataset/WikiQACorpus/WikiQA-test.tsv",
"pre_trained_emb": "./dataset/Glove/glove.840B.300d.txt"
},
"file_with_col_header": true,
"add_start_end_for_seq": true,
"file_header": {
"question_id": 0,
"question_text": 1,
"document_id": 2,
"document_title": 3,
"passage_id": 4,
"passage_text": 5,
"label": 6
},
"model_inputs": {
"question": ["question_text"],
"passage": ["passage_text"]
},
"target": ["label"]
},
"outputs":{
"save_base_dir": "./models/wikiqa_arcii/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.wikiqa_arcii/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"optimizer": {
"name": "Adam",
"params": {
"lr": 0.001
}
},
"fixed_lengths": {
"question": 200,
"passage": 200
},
"lr_decay": 0.9,
"minimum_lr": 0.00005,
"epoch_start_lr_decay": 20,
"use_gpu": true,
"cpu_num_workers": 1,
"batch_size": 64,
"batch_num_to_show_results": 500,
"max_epoch": 10,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"conf": {
"word": {
"cols": ["question_text", "passage_text"],
"dim": 300,
"fix_weight": true
}
}
},
{
"layer_id": "s1_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["question"]
},
{
"layer_id": "s2_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["passage"]
},
{
"layer_id": "s1_conv_1",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME",
"remind_lengths": false
},
"inputs": ["s1_dropout"]
},
{
"layer_id": "s2_conv_1",
"layer": "Conv",
"conf": {
"window_size": 3,
"output_channel_num": 32,
"padding_type": "SAME",
"remind_lengths": false
},
"inputs": ["s2_dropout"]
},
{
"layer_id": "match",
"layer": "Expand_plus",
"conf": {
},
"inputs": ["s1_conv_1", "s2_conv_1"]
},
{
"layer_id": "conv2D_1",
"layer": "Conv2D",
"conf": {
"window_size": [3,3],
"output_channel_num": 32,
"padding_type": "SAME"
},
"inputs": ["match"]
},
{
"layer_id": "pool2D_1",
"layer": "Pooling2D",
"conf": {
"window_size": [2,2]
},
"inputs": ["conv2D_1"]
},
{
"layer_id": "conv2D_2",
"layer": "Conv2D",
"conf": {
"window_size": [3,3],
"output_channel_num": 32,
"padding_type": "SAME"
},
"inputs": ["pool2D_1"]
},
{
"layer_id": "pool2D_2",
"layer": "Pooling2D",
"conf": {
"window_size": [2,2]
},
"inputs": ["conv2D_2"]
},
{
"layer_id": "flatten",
"layer": "Flatten",
"conf": {
},
"inputs": ["pool2D_2"]
},
{
"layer_id": "dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["flatten"]
},
{
"layer_id": "mlp",
"layer": "Linear",
"conf": {
"hidden_dim": [64, 32],
"activation": "ReLU",
"batch_norm": true,
"last_hidden_activation": true
},
"inputs": ["dropout"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "ReLU",
"batch_norm": true,
"last_hidden_activation": false,
"last_hidden_softmax": true
},
"inputs": ["mlp"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"weight": [0.1,0.9],
"size_average": true
},
"inputs": ["output","label"]
}
]
},
"metrics": ["auc", "accuracy"]
}

Просмотреть файл

@ -57,7 +57,7 @@
"valid_times_per_epoch": 5,
"fixed_lengths":{
"question": 30,
"passage": 120
"passage": 200
}
},
"architecture":[
@ -92,7 +92,6 @@
"layer": "Interaction",
"conf": {
"dropout": 0.2,
"hidden_dim": 300,
"matching_type": "general"
},
"inputs": ["question_dropout", "passage_dropout"]

Просмотреть файл

@ -0,0 +1,119 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 88.50 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"sequence": 0,
"tag": 1
},
"model_inputs": {
"words": ["sequence"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.015,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 2,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["sequence"],
"dim": 100
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["words"]
},
{
"layer_id": "sentence_BiLSTM",
"layer": "BiLSTM",
"conf": {
"hidden_dim": 100,
"num_layers": 1
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "lstm_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["sentence_BiLSTM"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["lstm_dropout"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"size_average": false,
"ignore_index": 0
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,208 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 90.36 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_ccnn_wcnn/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_ccnn_wcnn/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.005,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"CNNCharEmbedding",
"dropout": 0.5,
"dim": 50,
"embedding_matrix_dim": 30,
"stride":1,
"window_size": 3,
"activation": null
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["words"]
},
{
"layer_id": "Linear1",
"layer": "Linear",
"conf": {
"hidden_dim": [200],
"activation": "Tanh",
"batch_norm": true,
"last_hidden_activation": true,
"last_hidden_softmax": false
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "Conv1",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Linear1"]
},
{
"layer_id": "Dropout1",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["Conv1"]
},
{
"layer_id": "Conv2",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout1"]
},
{
"layer_id": "Dropout2",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["Conv2"]
},
{
"layer_id": "Conv3",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout2"]
},
{
"layer_id": "Dropout3",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["Conv3"]
},
{
"layer_id": "Conv4",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout3"]
},
{
"layer_id": "Dropout4",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["Conv4"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["Dropout4"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"size_average": false,
"ignore_index": 0
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,185 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 90.36 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_wcnn_ccnn_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_wcnn_ccnn_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.005,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"CNNCharEmbedding",
"dropout": 0.5,
"dim": 50,
"embedding_matrix_dim": 30,
"stride":1,
"window_size": 3,
"activation": null
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["words"]
},
{
"layer_id": "Linear1",
"layer": "Linear",
"conf": {
"hidden_dim": [200],
"activation": "Tanh",
"batch_norm": false,
"last_hidden_activation": true,
"last_hidden_softmax": false
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "Conv1",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Linear1"]
},
{
"layer_id": "Conv2",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Conv1"]
},
{
"layer_id": "Conv3",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Conv2"]
},
{
"layer_id": "Conv4",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Conv3"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["Conv4"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,130 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 90.63 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_ccnn_wlstm/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_ccnn_wlstm/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.015,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"CNNCharEmbedding",
"dropout": 0.5,
"dim": 50,
"embedding_matrix_dim": 30,
"stride":1,
"window_size": 3,
"activation": null
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["words"]
},
{
"layer_id": "sentence_BiLSTM",
"layer": "BiLSTM",
"conf": {
"hidden_dim": 100,
"num_layers": 1
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "lstm_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["sentence_BiLSTM"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["lstm_dropout"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"size_average": false,
"ignore_index": 0
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,134 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for slot sequence task. It achieved a f1-score of 91.38 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_ccnn_wlstm_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_ccnn_wlstm_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.015,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"CNNCharEmbedding",
"dropout": 0.5,
"dim": 50,
"embedding_matrix_dim": 30,
"stride":1,
"window_size": 3,
"activation": null
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["words"]
},
{
"layer_id": "sentence_BiLSTM",
"layer": "BiLSTM",
"conf": {
"hidden_dim": 100,
"num_layers": 1
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "lstm_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.5
},
"inputs": ["sentence_BiLSTM"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["lstm_dropout"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,191 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 90.27 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_clstm_wcnn_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_clstm_wcnn_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.005,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"LSTMCharEmbedding",
"dropout": 0.5,
"dim": 50,
"embedding_matrix_dim": 30,
"bidirect_flag": true
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["words"]
},
{
"layer_id": "Linear1",
"layer": "Linear",
"conf": {
"hidden_dim": [200],
"activation": "Tanh",
"batch_norm": false,
"last_hidden_activation": true,
"last_hidden_softmax": false
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "Conv1",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Linear1"]
},
{
"layer_id": "Conv2",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv1"]
},
{
"layer_id": "Conv3",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv2"]
},
{
"layer_id": "Conv4",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv3"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["Conv4"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,132 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 90.83 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word", "word_char"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_clstm_wlstm_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_clstm_wlstm_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.015,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
},
"char":{
"cols":["word_char"],
"type":"LSTMCharEmbedding",
"dropout": 0.2,
"dim": 50,
"embedding_matrix_dim": 30,
"bidirect_flag": true
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["words"]
},
{
"layer_id": "sentence_BiLSTM",
"layer": "BiLSTM",
"conf": {
"hidden_dim": 100,
"num_layers": 1
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "lstm_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["sentence_BiLSTM"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["lstm_dropout"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -23,13 +23,13 @@
},
"outputs":{
"save_base_dir": "./models/slot_tagging_encoder_decoder/",
"model_name": "model_debug.nb",
"train_log_name": "train_debug.log",
"test_log_name": "test_debug.log",
"predict_log_name": "predict_debug.log",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict_debug.tsv",
"cache_dir": ".cache.atis/"
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_encoder_decoder/"
},
"training_params": {
"vocabulary": {

Просмотреть файл

@ -0,0 +1,198 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 88.51 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_wcnn/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_wcnn/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.005,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["words"]
},
{
"layer_id": "Linear1",
"layer": "Linear",
"conf": {
"hidden_dim": [200],
"activation": "Tanh",
"batch_norm": true,
"last_hidden_activation": true,
"last_hidden_softmax": false
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "Conv1",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Linear1"]
},
{
"layer_id": "Dropout1",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["Conv1"]
},
{
"layer_id": "Conv2",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout1"]
},
{
"layer_id": "Dropout2",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["Conv2"]
},
{
"layer_id": "Conv3",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout2"]
},
{
"layer_id": "Dropout3",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["Conv3"]
},
{
"layer_id": "Conv4",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME"
},
"inputs": ["Dropout3"]
},
{
"layer_id": "Dropout4",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["Conv4"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "Linear",
"conf": {
"hidden_dim": [-1],
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["Dropout4"]
}
],
"loss": {
"losses": [
{
"type": "CrossEntropyLoss",
"conf": {
"size_average": false,
"ignore_index": 0
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,183 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 88.72 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_wcnn_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_wcnn_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.005,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.2
},
"inputs": ["words"]
},
{
"layer_id": "Linear1",
"layer": "Linear",
"conf": {
"hidden_dim": [200],
"activation": "Tanh",
"batch_norm": false,
"last_hidden_activation": true,
"last_hidden_softmax": false
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "Conv1",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Linear1"]
},
{
"layer_id": "Conv2",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv1"]
},
{
"layer_id": "Conv3",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv2"]
},
{
"layer_id": "Conv4",
"layer": "Conv",
"conf": {
"stride": 1,
"window_size": 3,
"output_channel_num": 200,
"batch_norm": true,
"activation": "ReLU",
"padding_type": "SAME",
"dropout": 0.2,
"remind_lengths": true
},
"inputs": ["Conv3"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["Conv4"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -0,0 +1,124 @@
{
"license": "Copyright (c) Microsoft Corporation. All rights reserved. Licensed under the MIT license.",
"tool_version": "1.1.0",
"model_description": "This model is used for sequence tagging task. It achieved a f1-score of 89.34 on the dataset conll-2003",
"inputs": {
"use_cache": true,
"dataset_type": "sequence_tagging",
"tagging_scheme": "BIOES",
"data_paths": {
"train_data_path": "./dataset/slot_tagging/conll_2003/eng.train.tsv",
"valid_data_path": "./dataset/slot_tagging/conll_2003/eng.testa.tsv",
"test_data_path": "./dataset/slot_tagging/conll_2003/eng.testb.tsv",
"pre_trained_emb": "./dataset/GloVe/glove.6B.100d.txt"
},
"add_start_end_for_seq": false,
"file_header": {
"word": 0,
"tag": 1
},
"model_inputs": {
"words": ["word"]
},
"target": ["tag"]
},
"outputs":{
"save_base_dir": "./models/slot_tagging_wlstm_crf/",
"model_name": "model.nb",
"train_log_name": "train.log",
"test_log_name": "test.log",
"predict_log_name": "predict.log",
"predict_fields": ["prediction"],
"predict_output_name": "predict.tsv",
"cache_dir": ".cache.slot_tagging_wlstm_crf/"
},
"training_params": {
"vocabulary": {
"min_word_frequency": 1
},
"cpu_num_workers": 4,
"optimizer": {
"name": "SGD",
"params": {
"lr": 0.015,
"weight_decay": 1e-8
}
},
"lr_decay": 0.95,
"minimum_lr": 0.00001,
"epoch_start_lr_decay": 1,
"use_gpu": true,
"batch_size": 10,
"batch_num_to_show_results": 500,
"max_epoch": 100,
"valid_times_per_epoch": 1
},
"architecture":[
{
"layer": "Embedding",
"weight_on_gpu": true,
"conf": {
"word": {
"cols": ["word"],
"dim": 100
}
}
},
{
"layer_id": "emb_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.4
},
"inputs": ["words"]
},
{
"layer_id": "sentence_BiLSTM",
"layer": "BiLSTM",
"conf": {
"hidden_dim": 100,
"num_layers": 1
},
"inputs": ["emb_dropout"]
},
{
"layer_id": "lstm_dropout",
"layer": "Dropout",
"conf": {
"dropout": 0.4
},
"inputs": ["sentence_BiLSTM"]
},
{
"layer_id": "rep2tag",
"layer": "Linear",
"conf": {
"hidden_dim": "#target#",
"activation": "PReLU",
"batch_norm": false,
"last_hidden_activation": false,
"last_hidden_softmax": false
},
"inputs": ["lstm_dropout"]
},
{
"output_layer_flag": true,
"layer_id": "output",
"layer": "CRF",
"conf": {},
"inputs": ["rep2tag"]
}
],
"loss": {
"losses": [
{
"type": "CRFLoss",
"conf": {
},
"inputs": ["output","tag"]
}
]
},
"metrics": ["seq_tag_f1","seq_tag_accuracy"]
}

Просмотреть файл

@ -34,13 +34,46 @@ def main(params):
lm = LearningMachine('predict', conf, problem, vocab_info=None, initialize=False, use_gpu=conf.use_gpu)
lm.load_model(conf.previous_model_path)
if params.predict_mode == 'batch':
logging.info('Predicting %s with the model saved at %s' % (conf.predict_data_path, conf.previous_model_path))
if params.predict_mode == 'batch':
lm.predict(conf.predict_data_path, conf.predict_output_path, conf.predict_file_columns, conf.predict_fields)
logging.info("Predict done! The predict result: %s" % conf.predict_output_path)
elif params.predict_mode == 'interactive':
print('='*80)
task_type = str(ProblemTypes[problem.problem_type]).split('.')[1]
sample_format = list(conf.predict_file_columns.keys())
target_ = conf.conf['inputs'].get('target', None)
target_list = list(target_) if target_ else []
for single_element in sample_format[:]:
if single_element in target_list:
sample_format.remove(single_element)
predict_file_columns = {}
for index, single in enumerate(sample_format):
predict_file_columns[single] = index
print('Enabling Interactive Inference Mode for %s Task...' % (task_type.upper()))
print('%s Task Interactive. The sample format is <%s>' % (task_type.upper(), ', '.join(sample_format)))
case_cnt = 1
while True:
print('Case%d:' % case_cnt)
sample = []
for single in sample_format:
temp_ = input('\t%s: ' % single)
if temp_.lower() == 'exit':
exit(0)
sample.append(temp_)
sample = '\t'.join(sample)
result = lm.interactive([sample], predict_file_columns, conf.predict_fields, params.predict_mode)
print('\tInference result: %s' % result)
case_cnt += 1
else:
raise Exception('Predict mode support interactive|batch, get %s' % params.predict_mode)
if __name__ == "__main__":
parser = argparse.ArgumentParser(description='Prediction')
parser.add_argument("--conf_path", type=str, help="configuration path")
parser.add_argument("--predict_mode", type=str, default='batch', help='interactive|batch')
parser.add_argument("--predict_data_path", type=str, help='specify another predict data path, instead of the one defined in configuration file')
parser.add_argument("--previous_model_path", type=str, help='load model trained previously.')
parser.add_argument("--predict_output_path", type=str, help='specify another prediction output path, instead of conf[outputs][save_base_dir] + conf[outputs][predict_output_name] defined in configuration file')

Просмотреть файл

@ -12,9 +12,9 @@ nltk.download('stopwords', quiet=True)
from utils.BPEEncoder import BPEEncoder
import os
import pickle as pkl
from utils.common_utils import load_from_pkl, dump_to_pkl
from utils.common_utils import load_from_pkl, dump_to_pkl, load_from_json, dump_to_json, prepare_dir, md5
from settings import ProblemTypes
from settings import ProblemTypes, Setting as st
import math
from utils.ProcessorsScheduler import ProcessorsScheduler
@ -65,6 +65,11 @@ class Problem():
target_with_start, target_with_end, target_with_unk, target_with_pad, same_length = (False, ) * 5
with_bos_eos = False
if ProblemTypes[problem_type] == ProblemTypes.sequence_tagging:
target_with_start = False
target_with_end = False
target_with_unk = False
self.lowercase = lowercase
self.problem_type = problem_type
self.tagging_scheme = tagging_scheme
@ -107,12 +112,9 @@ class Problem():
else:
return None
def get_data_generator_from_file(self, data_path_list, file_with_col_header, chunk_size=1000000):
# NOTE: file_path is a list type
for single_path in data_path_list:
def get_data_generator_from_file(self, data_path, file_with_col_header, chunk_size=1000000):
data_list = list()
if single_path is not None:
with open(single_path, "r", encoding='utf-8') as f:
with open(data_path, "r", encoding='utf-8') as f:
if file_with_col_header:
f.readline()
for index, line in enumerate(f):
@ -162,6 +164,9 @@ class Problem():
line_split[i] = self.text_preprocessor.preprocess(line_split[i])
if col_index_types[i] == 'word':
if ProblemTypes[self.problem_type] == ProblemTypes.sequence_tagging:
token_list = line_split[i].split(" ")
else:
token_list = self.tokenizer.tokenize(line_split[i])
docs[col_index_types[i]].append(token_list)
if 'char' in docs:
@ -218,11 +223,11 @@ class Problem():
def build(self, data_path_list, file_columns, input_types, file_with_col_header, answer_column_name, word2vec_path=None, word_emb_dim=None,
format=None, file_type=None, involve_all_words=None, file_format="tsv", show_progress=True,
cpu_num_workers=-1, max_vocabulary=800000, word_frequency=3):
cpu_num_workers=-1, max_vocabulary=800000, word_frequency=3, max_building_lines=1000*1000):
"""
Args:
training_data_path:
data_path_list:
file_columns: {
"word1": 0,
"word2": 1,
@ -260,26 +265,16 @@ class Problem():
"""
# parameter check
if not word2vec_path:
word_emb_dim, format, file_type, involve_all_words = None, None, None, None
if 'bpe' in input_types:
try:
bpe_encoder = BPEEncoder(input_types['bpe']['bpe_path'])
except KeyError:
raise Exception('Please define a bpe path at the embedding layer.')
else:
bpe_encoder = None
bpe_encoder = self._check_bpe_encoder(input_types)
self.file_column_num = len(file_columns)
progress = self.get_data_generator_from_file(data_path_list, file_with_col_header)
for data_path in data_path_list:
if data_path:
progress = self.get_data_generator_from_file(data_path, file_with_col_header, chunk_size=max_building_lines)
preprocessed_data_generator= self.build_training_multi_processor(progress, cpu_num_workers, file_columns, input_types, answer_column_name, bpe_encoder=bpe_encoder)
# update symbol universe
total_cnt_legal, total_cnt_illegal = 0, 0
for docs, target_docs, cnt_legal, cnt_illegal in tqdm(preprocessed_data_generator):
total_cnt_legal += cnt_legal
total_cnt_illegal += cnt_illegal
docs, target_docs, cnt_legal, cnt_illegal = next(preprocessed_data_generator)
# input_type
for input_type in input_types:
@ -292,7 +287,7 @@ class Problem():
elif ProblemTypes[self.problem_type] == ProblemTypes.regression or \
ProblemTypes[self.problem_type] == ProblemTypes.mrc:
pass
logging.info("Corpus imported: %d legal lines, %d illegal lines." % (total_cnt_legal, total_cnt_illegal))
logging.info("[Building Dictionary] in %s at most %d lines imported: %d legal lines, %d illegal lines." % (data_path, max_building_lines, cnt_legal, cnt_illegal))
# build dictionary
for input_type in input_types:
@ -300,6 +295,11 @@ class Problem():
logging.info("%d types in %s column" % (self.input_dicts[input_type].cell_num(), input_type))
if self.output_dict:
self.output_dict.build(threshold=0)
if ProblemTypes[self.problem_type] == ProblemTypes.sequence_tagging:
self.output_dict.cell_id_map["<start>"] = len(self.output_dict.cell_id_map)
self.output_dict.id_cell_map[len(self.output_dict.id_cell_map)] = "<start>"
self.output_dict.cell_id_map["<eos>"] = len(self.output_dict.cell_id_map)
self.output_dict.id_cell_map[len(self.output_dict.id_cell_map)] = "<eos>"
logging.info("%d types in target column" % (self.output_dict.cell_num()))
logging.debug("training data dict built")
@ -313,7 +313,12 @@ class Problem():
self.input_dicts['word'].update([list(word_emb_dict.keys())])
self.input_dicts['word'].build(threshold=0, max_vocabulary_num=len(word_emb_dict))
else:
word_emb_dict = load_embedding(word2vec_path, word_emb_dim, format, file_type, with_head=False, word_set=self.input_dicts['word'].cell_id_map.keys())
extend_vocabulary = set()
for single_word in self.input_dicts['word'].cell_id_map.keys():
extend_vocabulary.add(single_word)
if single_word.lower() != single_word:
extend_vocabulary.add(single_word.lower())
word_emb_dict = load_embedding(word2vec_path, word_emb_dim, format, file_type, with_head=False, word_set=extend_vocabulary)
for word in word_emb_dict:
loaded_emb_dim = len(word_emb_dict[word])
@ -329,11 +334,15 @@ class Problem():
word_emb_matrix = []
unknown_word_count = 0
scale = np.sqrt(3.0 / word_emb_dim)
for i in range(self.input_dicts['word'].cell_num()):
if self.input_dicts['word'].id_cell_map[i] in word_emb_dict:
word_emb_matrix.append(word_emb_dict[self.input_dicts['word'].id_cell_map[i]])
single_word = self.input_dicts['word'].id_cell_map[i]
if single_word in word_emb_dict:
word_emb_matrix.append(word_emb_dict[single_word])
elif single_word.lower() in word_emb_dict:
word_emb_matrix.append(word_emb_dict[single_word.lower()])
else:
word_emb_matrix.append(word_emb_dict['<unk>'])
word_emb_matrix.append(np.random.uniform(-scale, scale, word_emb_dim))
unknown_word_count += 1
word_emb_matrix = np.array(word_emb_matrix)
logging.info("word embedding matrix shape:(%d, %d); unknown word count: %d;" %
@ -382,8 +391,6 @@ class Problem():
def encode_data_multi_processor(self, data_generator, cpu_num_workers, file_columns, input_types, object_inputs,
answer_column_name, min_sentence_len, extra_feature, max_lengths=None, fixed_lengths=None, file_format="tsv", bpe_encoder=None):
for data in data_generator:
scheduler = ProcessorsScheduler(cpu_num_workers)
func_args = (data, file_columns, input_types, object_inputs,
@ -403,7 +410,7 @@ class Problem():
yield output_data, lengths, target, cnt_legal, cnt_illegal
def encode_data_list(self, data_list, file_columns, input_types, object_inputs, answer_column_name, min_sentence_len,
extra_feature, max_lengths=None, fixed_lengths=None, file_format="tsv", bpe_encoder=None):
extra_feature, max_lengths=None, fixed_lengths=None, file_format="tsv", bpe_encoder=None, predict_mode='batch'):
data = dict()
lengths = dict()
char_emb = True if 'char' in [single_input_type.lower() for single_input_type in input_types] else False
@ -423,6 +430,9 @@ class Problem():
type_branches = dict() # branch of input type, e.g. type_branches['query_index'] = 'query'
# for char: don't split these word
word_no_split = ['<start>', '<pad>', '<eos>', '<unk>']
for branch in object_inputs:
data[branch] = dict()
lengths[branch] = dict()
@ -461,11 +471,14 @@ class Problem():
line_split = line.rstrip().split('\t')
cnt_all += 1
if len(line_split) != len(file_columns):
# logging.warning("Current line is inconsistent with configuration/inputs/file_header. Ingore now. %s" % line)
if predict_mode == 'batch':
cnt_illegal += 1
if cnt_illegal / cnt_all > 0.33:
raise PreprocessError('The illegal data is too much. Please check the number of data columns or text token version.')
continue
else:
print('\tThe case is illegal! Please check your case and input again!')
return [None]*5
# cnt_legal += 1
length_appended_set = set() # to store branches whose length have been appended to lengths[branch]
@ -496,7 +509,7 @@ class Problem():
data[extra_info_type]['extra_passage_text'].append(line_split[i])
data[extra_info_type]['extra_passage_token_offsets'].append(passage_token_offsets)
else:
if extra_feature == False:
if extra_feature == False and ProblemTypes[self.problem_type] != ProblemTypes.sequence_tagging:
tokens = self.tokenizer.tokenize(line_split[i])
else:
tokens = line_split[i].split(' ')
@ -505,6 +518,28 @@ class Problem():
else:
tokens = line_split[i].split(' ')
# for sequence labeling task, the length must be record the corpus truth length
if ProblemTypes[self.problem_type] == ProblemTypes.sequence_tagging:
if not branch in length_appended_set:
lengths[branch]['sentence_length'].append(len(tokens))
length_appended_set.add(branch)
else:
if len(tokens) != lengths[branch]['sentence_length'][-1]:
# logging.warning(
# "The length of inputs are not consistent. Ingore now. %s" % line)
cnt_illegal += 1
if cnt_illegal / cnt_all > 0.33:
raise PreprocessError(
"The illegal data is too much. Please check the number of data columns or text token version.")
lengths[branch]['sentence_length'].pop()
true_len = len(lengths[branch]['sentence_length'])
# need delete the last example
check_list = ['data', 'lengths', 'target']
for single_check in check_list:
single_check = eval(single_check)
self.delete_example(single_check, true_len)
break
if fixed_lengths and type_branches[input_type[0]] in fixed_lengths:
if len(tokens) >= fixed_lengths[type_branches[input_type[0]]]:
tokens = tokens[:fixed_lengths[type_branches[input_type[0]]]]
@ -520,6 +555,8 @@ class Problem():
if self.with_bos_eos is True:
tokens = ['<start>'] + tokens + ['<eos>'] # so that source_with_start && source_with_end should be True
# for other tasks, length must be same as data length because fix/max_length operation
if not ProblemTypes[self.problem_type] == ProblemTypes.sequence_tagging:
if not branch in length_appended_set:
lengths[branch]['sentence_length'].append(len(tokens))
length_appended_set.add(branch)
@ -529,7 +566,8 @@ class Problem():
# "The length of inputs are not consistent. Ingore now. %s" % line)
cnt_illegal += 1
if cnt_illegal / cnt_all > 0.33:
raise PreprocessError("The illegal data is too much. Please check the number of data columns or text token version.")
raise PreprocessError(
"The illegal data is too much. Please check the number of data columns or text token version.")
lengths[branch]['sentence_length'].pop()
true_len = len(lengths[branch]['sentence_length'])
# need delete the last example
@ -544,8 +582,18 @@ class Problem():
temp_word_char = []
temp_word_length = []
for single_token in tokens:
temp_word_char.append(self.input_dicts[type2cluster[single_input_type]].lookup(single_token))
temp_word_length.append(len(single_token))
if single_token in word_no_split:
# temp_word_length.append(1)
temp_id = [self.input_dicts[type2cluster[single_input_type]].id(single_token)]
else:
temp_id = self.input_dicts[type2cluster[single_input_type]].lookup(single_token)
if fixed_lengths and 'word' in fixed_lengths:
if len(temp_id) >= fixed_lengths['word']:
temp_id = temp_id[:fixed_lengths['word']]
else:
temp_id = temp_id + [self.input_dicts[type2cluster[single_input_type]].id('<pad>')] * (fixed_lengths['word'] - len(temp_id))
temp_word_char.append(temp_id)
temp_word_length.append(len(temp_id))
data[branch][single_input_type].append(temp_word_char)
lengths[branch]['word_length'].append(temp_word_length)
else:
@ -625,7 +673,7 @@ class Problem():
def encode(self, data_path, file_columns, input_types, file_with_col_header, object_inputs, answer_column_name,
min_sentence_len, extra_feature, max_lengths=None, fixed_lengths=None, file_format="tsv", show_progress=True,
cpu_num_workers = -1):
cpu_num_workers=-1, chunk_size=1000*1000):
"""
Args:
@ -701,22 +749,16 @@ class Problem():
target: [...]
"""
if 'bpe' in input_types:
try:
bpe_encoder = BPEEncoder(input_types['bpe']['bpe_path'])
except KeyError:
raise Exception('Please define a bpe path at the embedding layer.')
else:
bpe_encoder = None
bpe_encoder = self._check_bpe_encoder(input_types)
progress = self.get_data_generator_from_file([data_path], file_with_col_header)
encoder_generator = self.encode_data_multi_processor(progress, cpu_num_workers,
progress = self.get_data_generator_from_file(data_path, file_with_col_header, chunk_size=chunk_size)
encode_generator = self.encode_data_multi_processor(progress, cpu_num_workers,
file_columns, input_types, object_inputs, answer_column_name, min_sentence_len, extra_feature, max_lengths,
fixed_lengths, file_format, bpe_encoder=bpe_encoder)
data, lengths, target = dict(), dict(), dict()
cnt_legal, cnt_illegal = 0, 0
for temp_data, temp_lengths, temp_target, temp_cnt_legal, temp_cnt_illegal in tqdm(encoder_generator):
for temp_data, temp_lengths, temp_target, temp_cnt_legal, temp_cnt_illegal in tqdm(encode_generator):
data = self._merge_encode_data(data, temp_data)
lengths = self._merge_encode_lengths(lengths, temp_lengths)
target = self._merge_target(target, temp_target)
@ -726,6 +768,59 @@ class Problem():
logging.info("%s: %d legal samples, %d illegal samples" % (data_path, cnt_legal, cnt_illegal))
return data, lengths, target
def build_encode_cache(self, conf, file_format="tsv"):
logging.info("[Cache] building encoding cache")
build_encode_cache_generator = self.get_encode_generator(conf, build_cache=True, file_format=file_format)
for _ in build_encode_cache_generator:
continue
logging.info("[Cache] encoding is saved to %s" % conf.encoding_cache_dir)
def get_encode_generator(self, conf, build_cache=True, file_format="tsv"):
# parameter check
if build_cache:
assert conf.encoding_cache_dir, 'There is no property encoding_cache_dir in object conf'
assert conf.encoding_cache_index_file_path, 'There is no property encoding_cache_index_file_path in object conf'
assert conf.encoding_cache_index_file_md5_path, 'There is no property encoding_cache_index_file_md5_path in object conf'
bpe_encoder = self._check_bpe_encoder(conf.input_types)
data_generator = self.get_data_generator_from_file(conf.train_data_path, conf.file_with_col_header, chunk_size=conf.chunk_size)
encode_generator = self.encode_data_multi_processor(data_generator, conf.cpu_num_workers,
conf.file_columns, conf.input_types, conf.object_inputs, conf.answer_column_name,
conf.min_sentence_len, conf.extra_feature, conf.max_lengths,
conf.fixed_lengths, file_format, bpe_encoder=bpe_encoder)
file_index = []
total_cnt_legal, total_cnt_illegal = 0, 0
for part_number, encode_data in enumerate(encode_generator):
data, lengths, target, cnt_legal, cnt_illegal = encode_data
if build_cache:
total_cnt_legal = total_cnt_legal + cnt_legal
total_cnt_illegal = total_cnt_illegal + cnt_illegal
file_name = st.cencoding_file_name_pattern % (part_number)
file_path = os.path.join(conf.encoding_cache_dir, file_name)
dump_to_pkl((data, lengths, target), file_path)
file_index.append([file_name, md5([file_path])])
logging.info("Up to now, in %s: %d legal samples, %d illegal samples" % (conf.train_data_path, total_cnt_legal, total_cnt_illegal))
yield data, lengths, target
if build_cache:
cache_index = dict()
cache_index[st.cencoding_key_index] = file_index
cache_index[st.cencoding_key_legal_cnt] = total_cnt_legal
cache_index[st.cencoding_key_illegal_cnt] = total_cnt_illegal
dump_to_json(cache_index, conf.encoding_cache_index_file_path)
dump_to_json(md5([conf.encoding_cache_index_file_path]), conf.encoding_cache_index_file_md5_path)
@staticmethod
def _check_bpe_encoder(input_types):
bpe_encoder = None
if 'bpe' in input_types:
try:
bpe_encoder = BPEEncoder(input_types['bpe']['bpe_path'])
except KeyError:
raise Exception('Please define a bpe path at the embedding layer.')
return bpe_encoder
def decode(self, model_output, lengths=None, batch_data=None):
""" decode the model output, either a batch of output or a single output

63
register_block.py Normal file
Просмотреть файл

@ -0,0 +1,63 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import os
import argparse
def get_block_path(block_name, path='./block_zoo'):
''' find the block_name.py file in block_zoo
Args:
block_name: the name need to be registered. eg. BiLSTM/ CRF
'''
get_dir = os.listdir(path)
for single in get_dir:
sub_dir = os.path.join(path, single)
if os.path.isdir(sub_dir):
result = get_block_path(block_name, path=sub_dir)
if result:
return result
else:
if block_name + '.py' == single:
return sub_dir
return None
def write_file(new_block_path, file_path):
init_path = os.path.join(file_path, '__init__.py')
diff = new_block_path[len(file_path):].split('/')
if diff[0] == '':
diff.pop(0)
# delete '.py' in the last str
diff[-1] = diff[-1][:-3]
line = 'from .' + diff[0] + ' import ' + diff[-1] + ', ' + diff[-1] + 'Conf'
with open(init_path, 'a', encoding='utf-8') as fin:
fin.write('\n' + line + '\n')
def register(block_name, new_block_path):
''' Add import code in the corresponding file. eg. block_zoo/__init__.py or block_zoo/subdir/__init__.py
'''
# check if block exist or not
if new_block_path:
block_path_split = new_block_path.split('/')
for i in range(len(block_path_split)-1, 1, -1):
# need_add_file.append(os.path.join('/'.join(block_path_split[:i])))
write_file(new_block_path, os.path.join('/'.join(block_path_split[:i])))
print('The block %s is registered successfully.' % block_name)
else:
raise Exception('The %s.py file does not exist! Please check your program or file name.' % block_name)
def main(params):
new_block_path = get_block_path(params.block_name)
register(params.block_name, new_block_path)
if __name__ == '__main__':
parse = argparse.ArgumentParser(description='Register Block')
parse.add_argument("--block_name", type=str, help="block name want to be registered")
params, _ = parse.parse_known_args()
assert params.block_name, 'Please specify a block_name via --block_name'
main(params)

Просмотреть файл

@ -18,11 +18,11 @@ LanguageTypes = Enum('LanguageTypes', ('english', 'chinese'))
ProblemTypes = Enum('ProblemTypes', ('sequence_tagging', 'classification', 'regression', 'mrc'))
# Supported sequence tagging scheme
TaggingSchemes = Enum('TaggingSchemes', ('BIO'))
TaggingSchemes = Enum('TaggingSchemes', ('BIO', 'BIOES'))
# supported metrics
SupportedMetrics = {
ProblemTypes.sequence_tagging: set(['seq_tag_f1', 'accuracy']),
ProblemTypes.sequence_tagging: set(['seq_tag_f1', 'seq_tag_accuracy']),
ProblemTypes.classification: set(['auc', 'accuracy', 'f1', 'macro_f1', 'macro_precision', 'macro_recall', 'micro_f1', 'micro_precision', 'micro_recall', 'weighted_f1', 'weighted_precision', 'weighted_recall']),
# In addition, for auc in multi-type classification,
# if there is a type named 1, auc@1 means use 1 as the positive label
@ -53,3 +53,27 @@ DefaultPredictionFields = {
# nltk's models
nltk.data.path.append(os.path.join(os.getcwd(), 'dataset', 'nltk_data'))
class Constant(type):
def __setattr__(self, name, value):
raise AttributeError("Class %s can not be modified"%(self.__name__))
class ConstantStatic(metaclass=Constant):
def __init__(self, *args,**kwargs):
raise Exception("Class %s can not be instantiated"%(self.__class__.__name__))
class Setting(ConstantStatic):
# cache
## cencoding (cache_encoding)
cencodig_index_file_name = 'index.json'
cencoding_index_md5_file_name = 'index_md5.json'
cencoding_file_name_pattern = 'encoding_cache_%s.pkl'
cencoding_key_finish = 'finish'
cencoding_key_index = 'index'
cencoding_key_legal_cnt = 'legal_line_cnt'
cencoding_key_illegal_cnt = 'illegal_line_cnt'

Просмотреть файл

Просмотреть файл

@ -0,0 +1,112 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
import sys
def BIO2BIOES(input_labels_list):
output_labels_list = []
for labels in input_labels_list:
new_labels = []
sent_len = len(labels)
for idx in range(sent_len):
if "-" not in labels[idx]:
new_labels.append(labels[idx])
else:
label_type = labels[idx].split('-')[-1]
if "B-" in labels[idx]:
if (idx == sent_len - 1) or ("I-" not in labels[idx + 1]):
new_labels.append("S-"+label_type)
else:
new_labels.append("B-"+label_type)
elif "I-" in labels[idx]:
if (idx == sent_len - 1) or ("I-" not in labels[idx + 1]):
new_labels.append("E-"+label_type)
else:
new_labels.append("I-"+label_type)
assert len(labels) == len(new_labels)
output_labels_list.append(new_labels)
return output_labels_list
def BIOES2BIO(input_labels_list):
output_labels_list = []
for labels in input_labels_list:
new_labels = []
sent_len = len(labels)
for idx in range(sent_len):
if "-" not in labels[idx]:
new_labels.append(labels[idx])
else:
label_type = labels[idx].split('-')[-1]
if "E-" in labels[idx]:
new_labels.append("I-" + label_type)
elif "S-" in labels[idx]:
new_labels.append("B-" + label_type)
else:
new_labels.append(labels[idx])
assert len(labels) == len(new_labels)
output_labels_list.append(new_labels)
return output_labels_list
def IOB2BIO(input_labels_list):
output_labels_list = []
for labels in input_labels_list:
new_labels = []
sent_len = len(labels)
for idx in range(sent_len):
if "I-" in labels[idx]:
label_type = labels[idx].split('-')[-1]
if (idx == 0) or (labels[idx - 1] == "O") or (label_type != labels[idx - 1].split('-')[-1]):
new_labels.append("B-" + label_type)
else:
new_labels.append(labels[idx])
else:
new_labels.append(labels[idx])
assert len(labels) == len(new_labels)
output_labels_list.append(new_labels)
return output_labels_list
if __name__ == '__main__':
'''Convert NER tagging schemes among IOB/BIO/BIOES.
For example: if you want to convert the IOB tagging scheme to BIO, then you run as following:
python taggingSchemes_Converter.py IOB2BIO input_iob_file output_bio_file
Input data format is tsv format.
'''
input_file_name, output_file_name = sys.argv[2], sys.argv[3]
words_list, labels_list, new_labels_list = [], [], []
with open(input_file_name, 'r') as input_file:
for line in input_file:
item = line.rstrip().split('\t')
assert len(item) == 2
words, labels = item[0].split(' '), item[1].split(' ')
if len(words) != len(labels):
print("Error line: " + line.rstrip())
continue
words_list.append(words)
labels_list.append(labels)
if sys.argv[1].upper() == "IOB2BIO":
print("Convert IOB -> BIO...")
new_labels_list = IOB2BIO(labels_list)
elif sys.argv[1].upper() == "BIO2BIOES":
print("Convert BIO -> BIOES...")
new_labels_list = BIO2BIOES(labels_list)
elif sys.argv[1].upper() == "BIOES2BIO":
print("Convert BIOES -> BIO...")
new_labels_list = BIOES2BIO(labels_list)
elif sys.argv[1].upper() == "IOB2BIOES":
print("Convert IOB -> BIOES...")
tmp_labels_list = IOB2BIO(labels_list)
new_labels_list = BIO2BIOES(tmp_labels_list)
else:
print("Argument error: sys.argv[1] should belongs to \"IOB2BIO/BIO2BIOES/BIOES2BIO/IOB2BIOES\"")
with open(output_file_name, 'w') as output_file:
for index in range(len(words_list)):
words, labels = words_list[index], new_labels_list[index]
line = " ".join(words) + '\t' + " ".join(labels) + '\n'
output_file.write(line)

110
train.py
Просмотреть файл

@ -1,7 +1,7 @@
# Copyright (c) Microsoft Corporation. All rights reserved.
# Licensed under the MIT license.
from settings import ProblemTypes, version
from settings import ProblemTypes, version, Setting as st
import os
import argparse
@ -15,7 +15,7 @@ import torch
import torch.nn as nn
from ModelConf import ModelConf
from problem import Problem
from utils.common_utils import dump_to_pkl, load_from_pkl, prepare_dir
from utils.common_utils import dump_to_pkl, load_from_pkl, load_from_json, dump_to_json, prepare_dir, md5
from utils.philly_utils import HDFSDirectTransferer
from losses import *
from optimizers import *
@ -34,6 +34,7 @@ class Cache:
self.dictionary_invalid = True
self.embedding_invalid = True
if not conf.pretrained_model_path:
# cache_conf
cache_conf = None
cache_conf_path = os.path.join(conf.cache_dir, 'conf_cache.json')
@ -57,10 +58,52 @@ class Cache:
self.embedding_invalid = False
self.dictionary_invalid = False
logging.info('[Cache] dictionary found')
return True
def _check_encoding(self, conf):
self.encoding_invalid = True
if not conf.pretrained_model_path and self.dictionary_invalid:
return False
# Calculate the MD5 of problem
problem_path = conf.problem_path if not conf.pretrained_model_path else conf.saved_problem_path
try:
conf.problem_md5 = md5([problem_path])
except Exception as e:
conf.problem_md5 = None
logging.info('Can not calculate md5 of problem.pkl from %s'%(problem_path))
return False
# check the valid of encoding cache
## encoding cache dir
conf.encoding_cache_dir = os.path.join(conf.cache_dir, conf.train_data_md5 + conf.problem_md5)
logging.debug('[Cache] conf.encoding_cache_dir %s' % (conf.encoding_cache_dir))
if not os.path.exists(conf.encoding_cache_dir):
return False
## encoding cache index
conf.encoding_cache_index_file_path = os.path.join(conf.encoding_cache_dir, st.cencodig_index_file_name)
conf.encoding_cache_index_file_md5_path = os.path.join(conf.encoding_cache_dir, st.cencoding_index_md5_file_name)
if not os.path.exists(conf.encoding_cache_index_file_path) or not os.path.exists(conf.encoding_cache_index_file_md5_path):
return False
if md5([conf.encoding_cache_index_file_path]) != load_from_json(conf.encoding_cache_index_file_md5_path):
return False
cache_index = load_from_json(conf.encoding_cache_index_file_path)
## encoding cache content
for index in cache_index[st.cencoding_key_index]:
file_name, file_md5 = index[0], index[1]
if file_md5 != md5([os.path.join(conf.encoding_cache_dir, file_name)]):
return False
if (st.cencoding_key_legal_cnt in cache_index) and (st.cencoding_key_illegal_cnt in cache_index):
conf.encoding_cache_legal_line_cnt = cache_index[st.cencoding_key_legal_cnt]
conf.encoding_cache_illegal_line_cnt = cache_index[st.cencoding_key_illegal_cnt]
self.encoding_invalid = False
logging.info('[Cache] encoding found')
logging.info('%s: %d legal samples, %d illegal samples' % (conf.train_data_path, conf.encoding_cache_legal_line_cnt, conf.encoding_cache_illegal_line_cnt))
return True
def check(self, conf, params):
@ -70,7 +113,7 @@ class Cache:
return
# encoding
if not self._check_encoding(conf):
self._renew_cache(params, conf.cache_dir)
self._renew_cache(params, conf.encoding_cache_dir)
def load(self, conf, problem, emb_matrix):
# load dictionary when (not finetune) and (cache valid)
@ -81,13 +124,17 @@ class Cache:
logging.info('[Cache] loading dictionary successfully')
if not self.encoding_invalid:
pass
self._prepare_encoding_cache(conf, problem, build=False)
logging.info('[Cache] preparing encoding successfully')
return problem, emb_matrix
def save(self, conf, params, problem, emb_matrix):
# make cache dir
if not os.path.exists(conf.cache_dir):
os.makedirs(conf.cache_dir)
shutil.copy(params.conf_path, os.path.join(conf.cache_dir, 'conf_cache.json'))
# dictionary
if self.dictionary_invalid:
if conf.mode == 'philly' and conf.emb_pkl_path.startswith('/hdfs/'):
with HDFSDirectTransferer(conf.problem_path, with_hdfs_command=True) as transferer:
@ -101,10 +148,11 @@ class Cache:
transferer.pkl_dump(emb_matrix)
else:
dump_to_pkl(emb_matrix, conf.emb_pkl_path)
logging.info("Embedding matrix saved to %s" % conf.emb_pkl_path)
logging.info("[Cache] Embedding matrix saved to %s" % conf.emb_pkl_path)
# encoding
if self.encoding_invalid:
pass
self._prepare_encoding_cache(conf, problem, build=params.make_cache_only)
def back_up(self, conf, problem):
cache_bakup_path = os.path.join(conf.save_base_dir, 'necessary_cache/')
@ -150,6 +198,34 @@ class Cache:
flag = False
return flag
def _prepare_encoding_cache(self, conf, problem, build=False):
# encoding cache dir
problem_path = conf.problem_path if not conf.pretrained_model_path else conf.saved_problem_path
conf.problem_md5 = md5([problem_path])
conf.encoding_cache_dir = os.path.join(conf.cache_dir, conf.train_data_md5 + conf.problem_md5)
if not os.path.exists(conf.encoding_cache_dir):
os.makedirs(conf.encoding_cache_dir)
# encoding cache files
conf.encoding_cache_index_file_path = os.path.join(conf.encoding_cache_dir, st.cencodig_index_file_name)
conf.encoding_cache_index_file_md5_path = os.path.join(conf.encoding_cache_dir, st.cencoding_index_md5_file_name)
conf.load_encoding_cache_generator = self._load_encoding_cache_generator
if build:
prepare_dir(conf.encoding_cache_dir, True, allow_overwrite=True, clear_dir_if_exist=True)
problem.build_encode_cache(conf)
self.encoding_invalid = False
if not self.encoding_invalid:
cache_index = load_from_json(conf.encoding_cache_index_file_path)
conf.encoding_file_index = cache_index[st.cencoding_key_index]
@staticmethod
def _load_encoding_cache_generator(cache_dir, file_index):
for index in file_index:
file_path = os.path.join(cache_dir, index[0])
yield load_from_pkl(file_path)
def main(params):
# init
conf = ModelConf("train", params.conf_path, version, params, mode=params.mode)
@ -172,6 +248,7 @@ def main(params):
# data preprocessing
## build dictionary when (not in finetune model) and (not use cache or cache invalid)
if (not conf.pretrained_model_path) and ((conf.use_cache == False) or cache.dictionary_invalid):
logging.info("="*100)
logging.info("Preprocessing... Depending on your corpus size, this step may take a while.")
# modify train_data_path to [train_data_path, valid_data_path, test_data_path]
# remember the test_data may be None
@ -181,11 +258,7 @@ def main(params):
word_emb_dim=conf.pretrained_emb_dim, format=conf.pretrained_emb_type,
file_type=conf.pretrained_emb_binary_or_text, involve_all_words=conf.involve_all_words_in_pretrained_emb,
show_progress=True if params.mode == 'normal' else False, cpu_num_workers = conf.cpu_num_workers,
max_vocabulary=conf.max_vocabulary, word_frequency=conf.min_word_frequency)
## encode rawdata when do not use cache
if conf.use_cache == False:
pass
max_vocabulary=conf.max_vocabulary, word_frequency=conf.min_word_frequency, max_building_lines=conf.max_building_lines)
# environment preparing
## cache save
@ -234,9 +307,20 @@ def main(params):
### optimizer
if isinstance(lm.model, nn.DataParallel):
optimizer = eval(conf.optimizer_name)(list(lm.model.parameters()) + list(lm.model.module.layers['embedding'].get_parameters()), **conf.optimizer_params)
if isinstance(lm.model.module.layers['embedding'].embeddings, nn.ModuleDict):
optimizer = eval(conf.optimizer_name)(list(lm.model.parameters()), **conf.optimizer_params)
else:
optimizer = eval(conf.optimizer_name)(list(lm.model.parameters()) + list(lm.model.layers['embedding'].get_parameters()), **conf.optimizer_params)
optimizer = eval(conf.optimizer_name)(
list(lm.model.parameters()) + list(lm.model.module.layers['embedding'].get_parameters()),
**conf.optimizer_params)
else:
if isinstance(lm.model.layers['embedding'].embeddings, nn.ModuleDict):
optimizer = eval(conf.optimizer_name)(
list(lm.model.parameters()), **conf.optimizer_params)
else:
optimizer = eval(conf.optimizer_name)(
list(lm.model.parameters()) + list(lm.model.layers['embedding'].get_parameters()),
**conf.optimizer_params)
## train
lm.train(optimizer, loss_fn)

Просмотреть файл

@ -3,6 +3,7 @@
import logging
import pickle as pkl
import json
import torch
import torch.nn as nn
import os
@ -49,6 +50,17 @@ def dump_to_pkl(obj, pkl_path):
pkl.dump(obj, fout, protocol=pkl.HIGHEST_PROTOCOL)
logging.debug("Obj dumped to %s!" % pkl_path)
def load_from_json(json_path):
data = None
with open(json_path, 'r', encoding='utf-8') as f:
data = json.loads(f.read())
logging.debug("%s loaded!" % json_path)
return data
def dump_to_json(obj, json_path):
with open(json_path, 'w', encoding='utf-8') as f:
f.write(json.dumps(obj))
logging.debug("Obj dumped to %s!" % json_path)
def get_trainable_param_num(model):
""" get the number of trainable parameters
@ -60,9 +72,15 @@ def get_trainable_param_num(model):
"""
if isinstance(model, nn.DataParallel):
if isinstance(model.module.layers['embedding'].embeddings, dict):
model_param = list(model.parameters()) + list(model.module.layers['embedding'].get_parameters())
else:
model_param = list(model.parameters())
else:
if isinstance(model.layers['embedding'].embeddings, dict):
model_param = list(model.parameters()) + list(model.layers['embedding'].get_parameters())
else:
model_param = list(model.parameters())
return sum(p.numel() for p in model_param if p.requires_grad)
@ -228,7 +246,7 @@ def md5(file_paths, chunk_size=1024*1024*1024):
""" Calculate a md5 of lists of files.
Args:
file_paths: an iterable object contains files. Files will be concatenated orderly if there are more than one file
file_paths: an iterable object contains file paths. Files will be concatenated orderly if there are more than one file
chunk_size: unit is byte, default value is 1GB
Returns:
md5
@ -243,3 +261,16 @@ def md5(file_paths, chunk_size=1024*1024*1024):
break
md5.update(data)
return md5.hexdigest()
def get_layer_class(model, layer_id):
"""get the layer class use layer_id
Args:
model: the model architecture, maybe nn.DataParallel type or model
layer_id: layer id from configuration
"""
if isinstance(model, nn.DataParallel):
return model.module.layers[layer_id]
else:
return model.layers[layer_id]

Просмотреть файл

@ -16,6 +16,7 @@ import codecs
import copy
from settings import ProblemTypes
import torch
import time
if sys.version_info < (3,):
@ -173,7 +174,7 @@ def corpus_permutation(*corpora):
return corpora_perm
def get_batches(problem, data, length, target, batch_size, input_types, pad_ids=None, permutate=False, transform_tensor=True):
def get_batches(problem, data, length, target, batch_size, input_types, pad_ids=None, permutate=False, transform_tensor=True, predict_mode='batch'):
"""
Args:
@ -232,13 +233,14 @@ def get_batches(problem, data, length, target, batch_size, input_types, pad_ids=
target_batches: ndarray/Variable shape: [number of batches, batch_size, targets]
"""
if predict_mode == 'batch':
logging.info("Start making batches")
if permutate is True:
#CAUTION! data and length would be revised
data = copy.deepcopy(data)
length = copy.deepcopy(length)
if target is not None:
target = copy.deepcopy(target)
# data = copy.deepcopy(data)
# length = copy.deepcopy(length)
# if target is not None:
# target = copy.deepcopy(target)
# shuffle the data
permutation = np.random.permutation(len(list(target.values())[0]))
@ -392,6 +394,7 @@ def get_batches(problem, data, length, target, batch_size, input_types, pad_ids=
target_batches.append(target_batch)
if predict_mode == 'batch':
logging.info("Batches got!")
return data_batches, length_batches, target_batches