Skip to content

Scope BertTokenClassifier task model and supporting infra #624

Open
@jbischof

Description

@jbischof

BertClassifier supports one output for the entire input sequence, but per-token classification is important for applications such as POS and NER tagging. This will require some scoping and design:

  1. Do we need a separate BertTokenClassifierPreprocessor or is just passing a label tensor the same length as the input enough?
  2. Do we want to offer a script that can turn a standard dataset like conll03 into something our preprocessor can use? In general most token labels are for "words" and not the subword tokens created by WordPiece/SentencePiece.
  3. Create a BertTokenClassifier task model with the correct task-specific layers and preprocessing.

Metadata

Metadata

Assignees

Labels

scoping requiredFeatures that need significant design and planning before being actionabletype:featureNew feature or request

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions