When preprocessing text data for an LLM fine-tuning task, why is it critical to apply subword tokenization (e. g., Byte-Pair Encoding) instead of word-based tokenization for handling rare or out-of-vocabulary words?
In the context of a natural language processing (NLP) application, which approach is most effective for implementing zero-shot learning to classify text data into categories that were not seen during training?
When deploying an LLM using NVIDIA Triton Inference Server for a real-time chatbot application, which optimization technique is most effective for reducing latency while maintaining high throughput?
Which principle of Trustworthy AI primarily concerns the ethical implications of AI's impact on society and includes considerations for both potential misuse and unintended consequences?