Google 《Prompt Engineering v7》
Writer Joey Haymaker Designer Michael Lanning Introduction 6 Prompt engineering 7 LLM output configuration 8 Output length 8 Sampling controls 9 Temperature 9 Top-K and top-P 10 Putting it all together AI or by using the API, because by prompting the model directly you will have access to the configuration such as temperature etc. This whitepaper discusses prompt engineering in detail. We will look configurations of a LLM. LLM output configuration Once you choose your model you will need to figure out the model configuration. Most LLMs come with various configuration options that control the LLM’s output0 码力 | 68 页 | 6.50 MB | 6 月前3DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model
vectors and the intermediate hidden states of routed experts) to ensure stable training. Under this configuration, DeepSeek-V2 comprises 236B total parameters, of which 21B are activated for each token. Training expert is 1408. Among the routed experts, 6 experts will be activated for each token. Under this configuration, DeepSeek-V2-Lite comprises 15.7B total parameters, of which 2.4B are activated for each token0 码力 | 52 页 | 1.23 MB | 1 年前3Facebook -- TVM AWS Meetup Talk
space (~10 lines of Relay IR) - A few days of work - TVM sampling model running in 30us on single server CPU core - Beat hand-written, highly optimized baselines (https://github.com/mozilla/LPCNet) by0 码力 | 11 页 | 3.08 MB | 5 月前3Deploy VTA on Intel FPGA
the compiled TVM to the SDCard Step 7: Install kernel module cma.ko and run apps/vta_rpc/start_rpc_server.sh Step 8: Configure vta/config/de10nano_config.json to vta_config.json Step 9: Go to vta/hardware/intel0 码力 | 12 页 | 1.35 MB | 5 月前3TVM Meetup: Quantization
Services, Inc. or its Affiliates. All rights reserved. Evaluation • Intel Cascade Lake 12-core Server • TFLite Pre-quantized Hosted Models© 2019, Amazon Web Services, Inc. or its Affiliates. All rights0 码力 | 19 页 | 489.50 KB | 5 月前3Trends Artificial Intelligence
intelligence. The earliest wave saw CapEx pouring into building internet infrastructure – massive server farms, undersea cables, and early data centers that enabled Amazon, Microsoft, Google and others AI Foundry expansion • NLWeb • Model Context Protocol (MCP) integration • Entra Agent ID • SQL Server 2025 • Windows Subsystem for Linux Open- Source • GitHub Copilot Chat Extension • Aurora AI-Powered0 码力 | 340 页 | 12.14 MB | 4 月前3
共 6 条
- 1