language-image pre-training grounded language
office the language DLL 'VBE7INTL.DLL' is not be found
其实是缺少vba,安装组件就可以了 解决方案:在卸载程序的面板中,找到office,右键选择更改,选择添加或删除功能,勾选如下两项就可以了 记得关掉office后再打开就好了 ......
学习笔记:A Survey on Large Language Model basedAutonomous Agents
挑选了自己感兴趣的部分整理了一下。 目录A Survey on Large Language Model basedAutonomous Agents1 LLM-AA Construction1.1 Architecture Design2 LLM-AA Application3 LLM-AA Ev ......
《PC Assembly Language》笔记——第一章-介绍
《PC Assembly Language》Paul A. Carter 第1章-介绍Introduction 1.2 Computer Organization 1.2.1 内存(Memory) 单位 大小 word 2 bytes double word 4 bytes quad word 8 ......
【论文阅读】Improving language understanding by generative pre-training
原始题目:Improving language understanding by generative pre-training 中文翻译:通过生成预训练提高语言理解能力 发表时间:2018年 平台:Preprint 文章链接:https://www.mikecaptain.com/resource ......
Language Models are Few-Shot Learners
郑重声明:原文参见标题,如有侵权,请联系作者,将会撤销发布! NeurIPS 2020 ......
SQL(Structured Query Language)简介和常见 SQL 命令示例
简介 SQL(Structured Query Language)是一种用于访问和操作关系型数据库的标准语言。它是一个功能强大的语言,用于执行各种数据库操作,包括检索数据、插入新记录、更新记录、删除记录、创建数据库、创建新表、设置权限以及执行存储过程和视图等。以下是 SQL 的一些重要方面: SQL ......
基于时间频率一致性对时间序列进行自监督对比预训练《Self-Supervised Contrastive Pre-Training for Time Series via Time-Frequency Consistency》(时序、时频一致性、对比学习)
2023年11月10日,今天看一篇论文,现在17:34,说实话,想摆烂休息,不想看,可还是要看,拴Q。 论文:Self-Supervised Contrastive Pre-Training for Time Series via Time-Frequency Consistency 或者是:Sel ......
一些WQL(WMI Query Language) 查询示例
目录WQL介绍一些WQL查询示例怎么执行WQL查询?WMIC在PowerShell里输入命令 WQL介绍 WQL (WMI Query Language) 是一种 SQL 的变体,用于查询和设置 Windows 管理工具(WMI,Windows Management Instrumentation) ......
TALLRec: An Effective and Efficient Tuning Framework to Align Large Language Model with Recommendation
目录概TallRec代码 Bao K., Zhang J., Zhang Y., Wang W., Feng F. and He X. TALLRec: An effective and efficient tuning framework to align large language model ......
【转】GN Language and Operation
原文链接:https://gn.googlesource.com/gn/+/refs/heads/main/docs/language.md 这里还有一篇:谷歌gn编译文件的使用简介 GN Language and Operation Contents GN Language and Operati ......
[论文阅读] EMO@ Earth Mover Distance Optimization for Auto-Regressive Language Modeling
Pre title: EMO: Earth Mover Distance Optimization for Auto-Regressive Language Modeling accepted: arXiv2023 paper: https://arxiv.org/abs/2310.04691 co ......
【论文阅读笔记】【Referring & Grounding】 Ferret: Refer and Ground Anything Anywhere at Any Granularity
读论文思考的问题 论文试图解决什么问题? referring 和 grounding 是图片空间信息理解的两大方面,而以往的模型通常只关注于其中的一个方面 能否根据人类理解图片的能力,将 referring 和 grounding 的能力统一到一个模型中? 如何联合这两个任务?它们能相互促进吗? 如 ......
多模态大模型的grounding能力
数据集 a)QW-VL:Visual Genome, RefCOCO, RefCOCO+, RefCOCOg, b)CogVLM:Visual7W,Flickr30K-Entities c)Kosmos2:GRIT OFA Unifying Architectures, Tasks, and Mod ......
GraphPrompt: Unifying Pre-Training and Downstream Tasks for Graph Neural Networks
目录概符号说明GraphPrompt代码 Liu Z., Yu X., Fang Y. and Zhang X. GraphPrompt: Unifying pre-training and downstream tasks for graph neural networks. WWW, 2023. ......
GPT-GNN: Generative Pre-Training of Graph Neural Networks
目录概符号说明GPT-GNN代码 Hu Z., Dong Y., Wang K., Chang K. and Sun Y. GPT-GNN: Generative pre-training of graph neural networks. KDD, 2020. 概 比较早的一篇图预训练模型. 符号 ......
Proj CDeepFuzz Paper Reading: POLYCRUISE: A Cross-Language Dynamic Information Flow Analysis
Abstract 本文: PolyCruise Method: 跨编程语言的holistic dynamic information flow analysis(DIFA) use a light language-specific analysis和language-agnostic online ......
Temporally Grounding Natural Sentence in Video
Temporally Grounding Natural Sentence in Video 摘要:我们引入了一种有效且高效的方法,可以在长的、未修剪的视频序列中建立(即本地化)自然句子。 具体来说,提出了一种新颖的时间地面网络(TGN)来暂时捕获视频和句子之间不断变化的细粒度逐字交互。 TGN 基 ......
论文阅读:Unifying Large Language Model and Knowledge Graph:A RoadMap
1 Introduction 大模型和知识图谱结合的综述。 简单介绍一下大模型和知识图谱的优缺点: 如上所示。 本文主要划分为三个模块,分别为: KG-enhanced LLMs LLM-augmented KGs Synergized LLM + KG 2 Background 主要介绍了LLM和 ......
Internet-augmented language models through few-shot prompting for open-domain question answering阅读笔记
Internet-augmented language models through few-shot prompting for open-domain question answering 其实我没怎么正经读过论文,尤其是带实验的,我目前认真读过的(大部头)也就是一些LLM的综述。记录这个文档主 ......
(待完善)C/C++ Language Standard
C89/C90 (ANSI C or ISO C) was the first standardized version of the language, released in 1989 and 1990, respectively C99 (ISO/IEC 9899:1999) C11 (ISO ......
Running Large Language Models locally – Your own ChatGPT-like AI in C#
For the past few months, a lot of news in tech as well as mainstream media has been around ChatGPT, an Artificial Intelligence (AI) product by the fol ......
TALL: Temporal Activity Localization via Language Query
1 introduction 确定任务:TALL(Temporal Activity Localization via Language):基于文本的时间活动定位,具体来说就是给定给定一个未修剪的视频和一个自然语言查询,目标是确定视频中所描述活动的开始和结束时间。 将视觉和文本特征嵌入到公共空间以获 ......
Top Programming Languages 2023
### 排名 ![image-20230915101511549](https://img2023.cnblogs.com/blog/540671/202309/540671-20230915103621203-1185888455.png) ### 工作 ![image-2023091510152 ......
Training language models to follow instructions with human feedback
郑重声明:原文参见标题,如有侵权,请联系作者,将会撤销发布! NeurIPS 2022 ......
《PROMPT2MODEL: Generating Deployable Models from Natural Language Instructions》论文学习
一、Introduction 传统上,从零开始构建一个自然语言处理(NLP)模型是一项重大任务。一个寻求解决新问题的NLP从业者需要定义他们的任务范围,找到或创建目标任务领域的行为数据,选择合适的模型架构,训练模型,通过评估评估其性能,然后将其部署到实际应用中。 Prompt2Model is a ......
Recommendation as Instruction Following: A Large Language Model Empowered Recommendation Approach
原文地址:https://arxiv.org/abs/2305.07001 本文作者将用户偏好、意图等构建为指令,并用这些指令调优一个LLM(3B Flan-T5-XL),该方法对用户友好,用户可以与系统交流获取更准确的推荐。 ## INTRODUCTION LLM是建立在自然语言文本上的,它不能直 ......
Proj CDeepFuzz Paper Reading: Software Testing with Large Language Model: Survey, Landscape, and Vision
## Abstract 本文: Task: Review on the use of LLMs in software testing Method: 1. analyzes 52 relevant studies ## 1. Intro ![](https://img2023.cnblogs.co ......
通过提示大语言模型进行个性化推荐LLM-Rec: Personalized Recommendation via Prompting Large Language Models
论文原文地址:https://arxiv.org/abs/2307.15780 本文提出了一种提示LLM并使用其生成的内容增强推荐系统的输入的方法,提高了个性化推荐的效果。 ## LLM-Rec Prompting ![](https://img2023.cnblogs.com/blog/17994 ......
Proj CDeepFuzz Paper Reading: Natural attack for pre-trained models of code
## Abstract 背景:目前大多数的adversarial attack method on pre-trained models of code忽略了perturbations should be natural to human judges(naturalness requirement ......
大模型时代的推荐系统Recommender Systems in the Era of Large Language Models (LLMs)
文章地址:https://arxiv.org/abs/2307.02046 笔记中的一些小实验中的模型都是基于GPT-3.5架构的ChatGPT模型。 本文主要讲述了比较具有代表性的方法利用LLM去学习user和item的表示,从预训练、微调和提示三个范式回顾了近期用于增强推荐系统的LLM先进技术, ......