COLLECTED BY
Organization:
Internet Archive
Focused crawls are collections of frequently-updated webcrawl data from narrow (as opposed to broad or wide) web crawls, often focused on a single domain or subdomain.
The Wayback Machine - https://web.archive.org/web/20210819110246/https://github.com/topics/nlu
Here are
416 public repositories
matching this topic...
💬 Open source machine learning framework to automate text- and voice-based conversations: NLU, dialogue management, connect to Slack, Facebook, and more - Create chatbots and voice assistants
Updated
Aug 19, 2021
Python
🤖 Dev tools to reliably understand text and automate conversations. Built-in NLU. Connect & deploy on any messaging channel (Slack, MS Teams, website, Telegram, etc).
Updated
Aug 19, 2021
TypeScript
Natural Language Processing Best Practices & Examples
Updated
Apr 8, 2021
Python
An NLP library for building bots, with entity extraction, sentiment analysis, automatic language identify, and so more
Updated
Aug 16, 2021
JavaScript
Snips Python library to extract meaning from text
Updated
May 3, 2021
Python
A model library for exploring state-of-the-art deep learning topologies and techniques for optimizing Natural Language Processing neural networks
Updated
Jun 10, 2021
Python
中文语言理解测评基准 Chinese Language Understanding Evaluation Benchmark: datasets, baselines, pre-trained models, corpus and leaderboard
Updated
Aug 3, 2021
Python
DELTA is a deep learning based natural language and speech processing platform.
Updated
Apr 16, 2021
Python
Updated
May 30, 2019
Clojure
The Open Source AI Chatbot Platform Builder in 100% C# Running in .NET Core with Machine Learning algorithm.
基于金融-司法领域(兼有闲聊性质)的聊天机器人,其中的主要模块有信息抽取、NLU、NLG、知识图谱等,并且利用Django整合了前端展示,目前已经封装了nlp和kg的restful接口
Rasa UI is a frontend for the Rasa Framework
Updated
Aug 4, 2021
JavaScript
User Simulation for Task-Completion Dialogues
Updated
Oct 5, 2017
OpenEdge ABL
🎯 🗯 Generate datasets for AI chatbots, NLP tasks, named entity recognition or text classification models using a simple DSL!
Updated
Jan 26, 2021
TypeScript
A platform for building conversational interfaces with intelligent agents (chatbots)
Updated
Aug 11, 2021
JavaScript
基于自然语言理解与机器学习的聊天机器人,支持多用户并发及自定义多轮对话
Updated
Jun 29, 2020
Python
Hardware-accelerated vector similarity search engine. Available as a HTTP service or as an embedded library.
An open platform for artificial intelligence, chat bots, virtual agents, social media automation, and live chat automation.
Updated
Jul 26, 2021
Java
JavaScript Web SDK for Dialogflow
Updated
Feb 20, 2019
TypeScript
1 line for thousands of State of The Art NLP models in hundreds of languages The fastest and most accurate way to solve text problems.
Updated
Jul 29, 2021
Python
API of Articut 中文斷詞 (兼具語意詞性標記):「斷詞」又稱「分詞」,是中文資訊處理的基礎。Articut 不用機器學習,不需資料模型,只用現代白話中文語法規則,即能達到 SIGHAN 2005 F1-measure 94% 以上,Recall 96% 以上的成績。
Updated
Jul 9, 2021
Python
A curated list of Open Information Extraction (OIE) resources: papers, code, data, etc.
Snips NLU rust implementation
Updated
Jan 28, 2020
Rust
all kinds of baseline models for sentence similarity 句子对语义相似度模型
Updated
Jul 5, 2018
Python
Recent Deep Learning papers in NLU and RL
Updated
Oct 5, 2019
Python
Updated
Jan 5, 2018
Jupyter Notebook
A collection of resources to make a smart speaker
Improve this page
Add a description, image, and links to the
nlu
topic page so that developers can more easily learn about it.
Curate this topic
Add this topic to your repo
To associate your repository with the
nlu
topic, visit your repo's landing page and select "manage topics."
Learn more
You can’t perform that action at this time.
You signed in with another tab or window. Reload to refresh your session.
You signed out in another tab or window. Reload to refresh your session.
Description
While using tokenizers.create with the model and vocab file for a custom corpus, the code throws an error and is not able to generate the BERT vocab file
Error Message
ValueError: Mismatch vocabulary! All special tokens specified must be control tokens in the sentencepiece vocabulary.
To Reproduce
from gluonnlp.data import tokenizers
tokenizers.create('spm', model_p