Initial RobertaTokenizer implementation (#365)
* Added initial RobertaTokenizer implementation * Added offset mapping to output * Updates for new custom op changes --------- Authored-by: Sayan Shaw <sayanshaw@microsoft.com>
This commit is contained in:
Родитель
7578af8361
Коммит
4d051b854b
|
@ -322,7 +322,7 @@ set(_HAS_TOKENIZER OFF)
|
|||
if(OCOS_ENABLE_GPT2_TOKENIZER)
|
||||
# GPT2
|
||||
set(_HAS_TOKENIZER ON)
|
||||
file(GLOB tok_TARGET_SRC "operators/tokenizer/gpt*.cc" "operators/tokenizer/unicode*.*" "operators/tokenizer/clip*.cc")
|
||||
file(GLOB tok_TARGET_SRC "operators/tokenizer/gpt*.cc" "operators/tokenizer/unicode*.*" "operators/tokenizer/clip*.cc" "operators/tokenizer/roberta*.cc")
|
||||
list(APPEND TARGET_SRC ${tok_TARGET_SRC})
|
||||
endif()
|
||||
|
||||
|
|
|
@ -20,6 +20,7 @@
|
|||
#include "nlohmann/json.hpp"
|
||||
#include "clip_tokenizer.hpp"
|
||||
#include "gpt2_tokenizer.hpp"
|
||||
#include "roberta_tokenizer.hpp"
|
||||
#include "string_tensor.h"
|
||||
#include "unicode.h"
|
||||
|
||||
|
|
|
@ -0,0 +1,247 @@
|
|||
// Copyright (c) Microsoft Corporation. All rights reserved.
|
||||
// Licensed under the MIT License.
|
||||
// Partial code comes from other Microsoft employee.
|
||||
|
||||
#include <string>
|
||||
#include <vector>
|
||||
#include <fstream>
|
||||
#include <sstream>
|
||||
#include <iostream>
|
||||
#include <algorithm>
|
||||
#include <list>
|
||||
#include <memory>
|
||||
#include <regex>
|
||||
#include <sstream>
|
||||
#include <stdexcept>
|
||||
#include <unordered_map>
|
||||
#include <functional>
|
||||
#include <codecvt>
|
||||
#include <mutex>
|
||||
|
||||
#include "nlohmann/json.hpp"
|
||||
#include "bpetokenizer.hpp"
|
||||
#include "string_tensor.h"
|
||||
#include "unicode.h"
|
||||
|
||||
// Note: the following logic comes from CPython: unicodetype_db.h (_PyUnicode_IsWhitespace)
|
||||
bool IsWithinUnicodeSpace(char32_t ch) {
|
||||
switch (ch) {
|
||||
case 0x0009:
|
||||
case 0x000A:
|
||||
case 0x000B:
|
||||
case 0x000C:
|
||||
case 0x000D:
|
||||
case 0x001C:
|
||||
case 0x001D:
|
||||
case 0x001E:
|
||||
case 0x001F:
|
||||
case 0x0020:
|
||||
case 0x0085:
|
||||
case 0x00A0:
|
||||
case 0x1680:
|
||||
case 0x2000:
|
||||
case 0x2001:
|
||||
case 0x2002:
|
||||
case 0x2003:
|
||||
case 0x2004:
|
||||
case 0x2005:
|
||||
case 0x2006:
|
||||
case 0x2007:
|
||||
case 0x2008:
|
||||
case 0x2009:
|
||||
case 0x200A:
|
||||
case 0x2028:
|
||||
case 0x2029:
|
||||
case 0x202F:
|
||||
case 0x205F:
|
||||
case 0x3000:
|
||||
return true;
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
bool IsEmptyuString(const ustring& str) {
|
||||
return std::all_of(str.begin(), str.end(), [](char32_t ch) { return IsWithinUnicodeSpace(ch); });
|
||||
}
|
||||
|
||||
KernelRobertaBpeTokenizer::KernelRobertaBpeTokenizer(const OrtApi& api, const OrtKernelInfo& info)
|
||||
: BaseKernel(api, info) {
|
||||
std::string vocab = ort_.KernelInfoGetAttribute<std::string>(&info, "vocab");
|
||||
if (vocab.empty()) {
|
||||
ORTX_CXX_API_THROW("vocabulary shouldn't be empty.", ORT_INVALID_ARGUMENT);
|
||||
}
|
||||
|
||||
std::string merges = ort_.KernelInfoGetAttribute<std::string>(&info, "merges");
|
||||
if (merges.empty()) {
|
||||
ORTX_CXX_API_THROW("merges shouldn't be empty.", ORT_INVALID_ARGUMENT);
|
||||
}
|
||||
|
||||
if (!TryToGetAttribute<int64_t>("padding_length", padding_length_)) {
|
||||
padding_length_ = -1;
|
||||
}
|
||||
|
||||
if (padding_length_ != -1 && padding_length_ <= 0) {
|
||||
ORTX_CXX_API_THROW("padding_length should be more than 0 or equal -1", ORT_INVALID_ARGUMENT);
|
||||
}
|
||||
|
||||
std::stringstream vocabu_stream(vocab);
|
||||
std::stringstream merges_stream(merges);
|
||||
bbpe_tokenizer_ = std::make_shared<VocabData>();
|
||||
bbpe_tokenizer_->Load(vocabu_stream, merges_stream, "<|endoftext|>", "<|endoftext|>");
|
||||
}
|
||||
|
||||
std::vector<int64_t> KernelRobertaBpeTokenizer::Tokenize(ustring& input, int64_t max_length, std::list<std::list<std::pair<int, int>>>& offset_map) {
|
||||
std::vector<int64_t> res;
|
||||
|
||||
if (IsEmptyuString(input)) {
|
||||
return res;
|
||||
}
|
||||
// Add BOS token to result
|
||||
res.push_back(bbpe_tokenizer_->GetEncoding("<s>"));
|
||||
|
||||
// Parse input
|
||||
auto special_token_split_res = bbpe_tokenizer_->SplitBySpecialTokens(input);
|
||||
TokenWithRegularExp regcmp;
|
||||
|
||||
for (auto& seg_id : special_token_split_res) {
|
||||
if (static_cast<int64_t>(res.size()) >= max_length) break;
|
||||
|
||||
if (seg_id.second != -1) {
|
||||
res.push_back(seg_id.second);
|
||||
continue;
|
||||
}
|
||||
|
||||
auto cur_input = std::move(seg_id.first);
|
||||
// Note: keep ptr to make sure the string_view is valid in the following process
|
||||
const char32_t* ptr = cur_input.c_str();
|
||||
regcmp.Set(ptr);
|
||||
|
||||
int offset = 0;
|
||||
std::list<std::pair<int, int>> offset_mapping;
|
||||
|
||||
// Add offset mapping for BOS token
|
||||
offset_mapping.push_back(std::make_pair(0, 0));
|
||||
|
||||
while (static_cast<int64_t>(res.size()) < max_length) {
|
||||
auto [b, tok] = regcmp.GetNextToken();
|
||||
if (!b) break;
|
||||
|
||||
std::string utf8_token = std::string(ustring(tok));
|
||||
|
||||
// Handle offset mapping and special cases
|
||||
if (utf8_token.at(0) == ' ') {
|
||||
offset_mapping.push_back(std::make_pair(offset + 1, offset + utf8_token.size()));
|
||||
} else {
|
||||
offset_mapping.push_back(std::make_pair(offset, offset + utf8_token.size()));
|
||||
}
|
||||
offset += utf8_token.size();
|
||||
|
||||
// Get byte encodings prior to performing BPE
|
||||
byte_list_.clear();
|
||||
for (char& cp : utf8_token) {
|
||||
byte_list_.push_back(bbpe_tokenizer_->ByteEncoder()[static_cast<unsigned char>(cp)]);
|
||||
}
|
||||
|
||||
// Perform BPE
|
||||
bbpe_tokenizer_->bpe(byte_list_);
|
||||
|
||||
// Add output to result
|
||||
for (auto p : byte_list_) {
|
||||
if (static_cast<int64_t>(res.size()) >= max_length) {
|
||||
break;
|
||||
}
|
||||
|
||||
res.push_back(p);
|
||||
}
|
||||
}
|
||||
// Add offset mapping for EOS token
|
||||
offset_mapping.push_back(std::make_pair(0, 0));
|
||||
|
||||
// Add offset mappings for input in this instance to list of offset mappings for all inputs
|
||||
offset_map.push_back(offset_mapping);
|
||||
}
|
||||
// Add EOS token to result
|
||||
res.push_back(bbpe_tokenizer_->GetEncoding("</s>"));
|
||||
return res;
|
||||
}
|
||||
|
||||
void KernelRobertaBpeTokenizer::Compute(OrtKernelContext* context) {
|
||||
// Setup inputs
|
||||
const OrtValue* input = ort_.KernelContext_GetInput(context, 0);
|
||||
std::vector<std::string> str_input;
|
||||
std::list<std::list<std::pair<int, int>>> offset_map;
|
||||
GetTensorMutableDataString(api_, ort_, context, input, str_input);
|
||||
OrtTensorDimensions input_dim(ort_, input);
|
||||
|
||||
std::vector<std::vector<int64_t>> tokenize_results;
|
||||
for (auto& str : str_input) {
|
||||
ustring ustr = ustring(str);
|
||||
tokenize_results.emplace_back(Tokenize(ustr, padding_length_ < 0 ? INT64_MAX : padding_length_, offset_map));
|
||||
}
|
||||
|
||||
size_t max_length = 0;
|
||||
if (padding_length_ == -1) {
|
||||
for (auto& res : tokenize_results) {
|
||||
max_length = std::max(max_length, res.size());
|
||||
}
|
||||
} else {
|
||||
max_length = static_cast<size_t>(padding_length_);
|
||||
}
|
||||
|
||||
OrtTensorDimensions output_dim = input_dim;
|
||||
output_dim.push_back(max_length);
|
||||
|
||||
OrtTensorDimensions offset_dim = output_dim;
|
||||
offset_dim.push_back(2); // tuple of offsets for each input id
|
||||
|
||||
OrtValue* tokenize_output = ort_.KernelContext_GetOutput(context, 0, output_dim.data(), output_dim.size());
|
||||
OrtValue* attention_mask = ort_.KernelContext_GetOutput(context, 1, output_dim.data(), output_dim.size());
|
||||
OrtValue* offset_mapping = ort_.KernelContext_GetOutput(context, 2, offset_dim.data(), offset_dim.size());
|
||||
auto* token = ort_.GetTensorMutableData<int64_t>(tokenize_output);
|
||||
auto* mask = ort_.GetTensorMutableData<int64_t>(attention_mask);
|
||||
auto* offset = ort_.GetTensorMutableData<int64_t>(offset_mapping);
|
||||
|
||||
int idx = 0;
|
||||
for (auto& res : tokenize_results) {
|
||||
for (int64_t id : res) {
|
||||
token[idx] = id;
|
||||
mask[idx] = 1;
|
||||
idx++;
|
||||
}
|
||||
|
||||
for (size_t i = res.size(); i < max_length; i++) {
|
||||
token[idx] = 0;
|
||||
mask[idx] = 0;
|
||||
idx++;
|
||||
}
|
||||
}
|
||||
|
||||
int idx2 = 0;
|
||||
for (auto& res : offset_map) {
|
||||
for (auto& mapping : res) {
|
||||
offset[idx2] = mapping.first;
|
||||
idx2++;
|
||||
offset[idx2] = mapping.second;
|
||||
idx2++;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
const char* CustomOpRobertaBpeTokenizer::GetName() const {
|
||||
return "RobertaTokenizer";
|
||||
}
|
||||
|
||||
size_t CustomOpRobertaBpeTokenizer::GetInputTypeCount() const {
|
||||
return 1;
|
||||
}
|
||||
|
||||
ONNXTensorElementDataType CustomOpRobertaBpeTokenizer::GetInputType(size_t /*index*/) const {
|
||||
return ONNX_TENSOR_ELEMENT_DATA_TYPE_STRING;
|
||||
}
|
||||
size_t CustomOpRobertaBpeTokenizer::GetOutputTypeCount() const {
|
||||
return 3;
|
||||
}
|
||||
|
||||
ONNXTensorElementDataType CustomOpRobertaBpeTokenizer::GetOutputType(size_t /*index*/) const {
|
||||
return ONNX_TENSOR_ELEMENT_DATA_TYPE_INT64;
|
||||
}
|
|
@ -0,0 +1,26 @@
|
|||
#include <list>
|
||||
#include "ocos.h"
|
||||
#include "ustring.h"
|
||||
#include "string_utils.h"
|
||||
|
||||
class VocabData;
|
||||
|
||||
struct KernelRobertaBpeTokenizer : BaseKernel {
|
||||
KernelRobertaBpeTokenizer(const OrtApi& api, const OrtKernelInfo& info);
|
||||
void Compute(OrtKernelContext* context);
|
||||
|
||||
private:
|
||||
std::vector<int64_t> Tokenize(ustring& input, int64_t max_length, std::list<std::list<std::pair<int, int>>>& offset_map);
|
||||
|
||||
int64_t padding_length_;
|
||||
std::list<int> byte_list_;
|
||||
std::shared_ptr<VocabData> bbpe_tokenizer_;
|
||||
};
|
||||
|
||||
struct CustomOpRobertaBpeTokenizer : OrtW::CustomOpBase<CustomOpRobertaBpeTokenizer, KernelRobertaBpeTokenizer> {
|
||||
const char* GetName() const;
|
||||
size_t GetInputTypeCount() const;
|
||||
ONNXTensorElementDataType GetInputType(size_t index) const;
|
||||
size_t GetOutputTypeCount() const;
|
||||
ONNXTensorElementDataType GetOutputType(size_t index) const;
|
||||
};
|
|
@ -4,6 +4,7 @@
|
|||
#ifdef ENABLE_GPT2_TOKENIZER
|
||||
#include "gpt2_tokenizer.hpp"
|
||||
#include "clip_tokenizer.hpp"
|
||||
#include "roberta_tokenizer.hpp"
|
||||
#endif
|
||||
|
||||
#ifdef ENABLE_SPM_TOKENIZER
|
||||
|
@ -31,6 +32,7 @@ FxLoadCustomOpFactory LoadCustomOpClasses_Tokenizer = LoadCustomOpClasses<
|
|||
#ifdef ENABLE_GPT2_TOKENIZER
|
||||
, CustomOpBpeTokenizer
|
||||
, CustomOpClipBpeTokenizer
|
||||
, CustomOpRobertaBpeTokenizer
|
||||
#endif
|
||||
|
||||
#ifdef ENABLE_SPM_TOKENIZER
|
||||
|
|
|
@ -0,0 +1,89 @@
|
|||
import unittest
|
||||
import numpy as np
|
||||
import numpy.lib.recfunctions as nlr
|
||||
import onnxruntime as _ort
|
||||
|
||||
from pathlib import Path
|
||||
from onnx import helper, onnx_pb as onnx_proto
|
||||
from transformers import RobertaTokenizerFast
|
||||
from onnxruntime_extensions import (
|
||||
make_onnx_model,
|
||||
get_library_path as _get_library_path)
|
||||
|
||||
def _get_file_content(path):
|
||||
with open(path, "rb") as file:
|
||||
return file.read()
|
||||
|
||||
|
||||
def _create_test_model(**kwargs):
|
||||
vocab_file = kwargs["vocab_file"]
|
||||
merges_file = kwargs["merges_file"]
|
||||
max_length = kwargs["max_length"]
|
||||
|
||||
node = [helper.make_node(
|
||||
'RobertaTokenizer', ['string_input'], ['input_ids', 'attention_mask', 'offset_mapping'], vocab=_get_file_content(vocab_file),
|
||||
merges=_get_file_content(merges_file), name='bpetok', padding_length=max_length,
|
||||
domain='ai.onnx.contrib')]
|
||||
input1 = helper.make_tensor_value_info(
|
||||
'string_input', onnx_proto.TensorProto.STRING, [None])
|
||||
output1 = helper.make_tensor_value_info(
|
||||
'input_ids', onnx_proto.TensorProto.INT64, ["batch_size", "num_input_ids"])
|
||||
output2 = helper.make_tensor_value_info(
|
||||
'attention_mask', onnx_proto.TensorProto.INT64, ["batch_size", "num_attention_masks"])
|
||||
output3 = helper.make_tensor_value_info(
|
||||
'offset_mapping', onnx_proto.TensorProto.INT64, ["batch_size", "num_offsets", 2])
|
||||
|
||||
graph = helper.make_graph(node, 'test0', [input1], [output1, output2, output3])
|
||||
model = make_onnx_model(graph)
|
||||
return model
|
||||
|
||||
|
||||
class TestRobertaTokenizer(unittest.TestCase):
|
||||
@classmethod
|
||||
def setUpClass(cls):
|
||||
cls.tokenizer = RobertaTokenizerFast.from_pretrained("roberta-base")
|
||||
files = cls.tokenizer.save_vocabulary(".")
|
||||
cls.tokjson = files[0]
|
||||
cls.merges = files[1]
|
||||
|
||||
def _run_tokenizer(self, test_sentence, padding_length=-1):
|
||||
model = _create_test_model(vocab_file=self.tokjson, merges_file=self.merges, max_length=padding_length)
|
||||
so = _ort.SessionOptions()
|
||||
so.register_custom_ops_library(_get_library_path())
|
||||
sess = _ort.InferenceSession(model.SerializeToString(), so)
|
||||
input_text = np.array(test_sentence)
|
||||
input_ids, attention_mask, offset_mapping = sess.run(None, {'string_input': input_text})
|
||||
print("\nTest Sentence: " + str(test_sentence))
|
||||
print("\nInput IDs: " + str(input_ids))
|
||||
print("Attention Mask: " + str(attention_mask))
|
||||
# Reformat offset mapping from 3d array to 2d array of tuples before printing for readability
|
||||
reformatted_offset_mapping = nlr.unstructured_to_structured(np.array(offset_mapping)).astype('O')
|
||||
print("Offset Mapping: " + str(reformatted_offset_mapping))
|
||||
roberta_out = self.tokenizer(test_sentence, return_offsets_mapping=True)
|
||||
expect_input_ids = roberta_out['input_ids']
|
||||
expect_attention_mask = roberta_out['attention_mask']
|
||||
expect_offset_mapping = roberta_out['offset_mapping']
|
||||
print("\nExpected Input IDs: " + str(expect_input_ids))
|
||||
print("Expected Attention Mask: " + str(expect_attention_mask))
|
||||
print("Expected Offset Mapping: " + str(expect_offset_mapping) + "\n")
|
||||
np.testing.assert_array_equal(expect_input_ids, input_ids)
|
||||
np.testing.assert_array_equal(expect_attention_mask, attention_mask)
|
||||
np.testing.assert_array_equal(expect_offset_mapping, offset_mapping)
|
||||
|
||||
del sess
|
||||
del so
|
||||
|
||||
def test_tokenizer(self):
|
||||
self._run_tokenizer(["I can feel the magic, can you?"])
|
||||
self._run_tokenizer(["Hey Cortana"])
|
||||
self._run_tokenizer(["lower newer"])
|
||||
self._run_tokenizer(["a diagram", "a dog", "a cat"])
|
||||
self._run_tokenizer(["a photo of a cat", "a photo of a dog"])
|
||||
self._run_tokenizer(["one + two = three"])
|
||||
self._run_tokenizer(["9 8 7 6 5 4 3 2 1 0"])
|
||||
self._run_tokenizer(["9 8 7 - 6 5 4 - 3 2 1 0"])
|
||||
self._run_tokenizer(["One Microsoft Way, Redmond, WA"])
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
unittest.main()
|
|
@ -1 +1 @@
|
|||
0.7.0
|
||||
0.7.0
|
Загрузка…
Ссылка в новой задаче