Yahoo Canada Web Search

  1. Ad

    related to: Is codegeex a good AI code generator?
  2. Cody uses a deep understanding of your codebase to explain, fix, and help you write code. Cody helps you code faster with AI-assisted auto-complete in any programming language.

Search results

      • CodeGeeX stands as a powerful AI-powered coding assistant, significantly enhancing productivity and learning for developers across the spectrum. Its seamless integration with the Hugging Face ecosystem, coupled with its robust feature set and community support, makes it an essential tool for anyone looking to streamline their coding process.
  1. People also ask

  2. 1 day ago · Selling Point of CodeGeeX: Ideal AI-code generator for cross-platform development. This free AI Code Generator is a multilingual code generation tool trained on 13 billion parameters and 20 programming languages, including Java, JavaScript, C++, and Python.

    • Goodfirms
    • Is codegeex a good AI code generator?1
    • Is codegeex a good AI code generator?2
    • Is codegeex a good AI code generator?3
    • Is codegeex a good AI code generator?4
    • Is codegeex a good AI code generator?5
    • Overview
    • CodeGeeX: A Multilingual Code Generation Model
    • News
    • Getting Started
    • CodeGeeX: Architecture, Code Corpus, and Implementation
    • HumanEval-X: A new benchmark for Multilingual Program Synthesis
    • License
    • Citation

    •CodeGeeX: A Multilingual Code Generation Model

    •News

    •Getting Started

    •Installation

    •Model Weights

    •Inference on GPUs

    We introduce CodeGeeX, a large-scale multilingual code generation model with 13 billion parameters, pre-trained on a large code corpus of more than 20 programming languages. As of June 22, 2022, CodeGeeX has been trained on more than 850 billion tokens on a cluster of 1,536 Ascend 910 AI Processors. CodeGeeX has several unique features:

    •Multilingual Code Generation: CodeGeeX has good performance for generating executable programs in several mainstream programming languages, including Python, C++, Java, JavaScript, Go, etc. DEMO

    •Crosslingual Code Translation: CodeGeeX supports the translation of code snippets between different languages. Simply by one click, CodeGeeX can transform a program into any expected language with a high accuracy. DEMO

    •Customizable Programming Assistant: CodeGeeX is available in the VS Code extension marketplace for free. It supports code completion, explanation, summarization and more, which empower users with a better coding experience. VS Code Extension

    •Open-Source and Cross-Platform: All codes and model weights are publicly available for research purposes. CodeGeeX supports both Ascend and NVIDIA platforms. It supports inference in a single Ascend 910, NVIDIA V100 or A100. Apply Model Weights

    HumanEval-X for Realistic Multilingual Benchmarking. To help standardize the evaluation of multilingual code generation and translation, we develop and release the HumanEval-X Benchmark. HumanEval-X is a new multilingual benchmark that contains 820 human-crafted coding problems in 5 programming languages (Python, C++, Java, JavaScript, and Go), each of these problems is associated with tests and solutions. Usage 🤗 Available in HuggingFace

    •🌟 2023-07-24: CodeGeeX2 has been released, more powerful, faster, and lightweight. Support 100+ languages and many new features.

    •2023-5-16: CodeGeeX paper has been accepted by KDD 2023, Long Beach and will be represented during the conference.

    •2023-03-30: CodeGeeX paper is now available at arxiv.

    •2023-02-14: CodeGeeX now supports Cloud Studio, a fantastic web IDE from Tencent. Click on the badge on top of this page to quickly launch an environment to test CodeGeeX.

    •2023-02-13: Thanks a lot to OneFlow team for adding oneflow backend for CodeGeeX's inference (Even faster than FasterTransformer under FP16!). Check more details here.

    •2023-02: We are hosting CodeGeeX "Coding With AI" Hackathon, design cool applications based on CodeGeeX and win prizes (RTX 4090, DJI drone, etc)!

    Installation

    Python 3.7+ / CUDA 11+ / PyTorch 1.10+ / DeepSpeed 0.6+ are required. Install codegeex package via: Or use CodeGeeX docker to quickly set up the environment (with nvidia-docker installed):

    Model Weights

    Apply and download model weights through this link. You'll receive by mail urls.txt that contains temporary download links. We recommend you to use aria2 to download it via the following command (Please make sure you have enough disk space to download the checkpoint (~26GB)): Run the following command to get the full model weights:

    Inference on GPUs

    Have a try on generating the first program with CodeGeeX. First, specify the path of the model weights in configs/codegeex_13b.sh. Second, write the prompt (natural language description or code snippet) into a file, e.g., tests/test_prompt.txt, then run the following script:

    Architecture: CodeGeeX is a large-scale pre-trained programming language model based on transformers. It is a left-to-right autoregressive decoder, which takes code and natural language as input and predicts the probability of the next token. CodeGeeX contains 40 transformer layers with a hidden size of 5,120 for self-attention blocks and 20,480 for feed-forward layers, making its size reach 13 billion parameters. It supports a maximum sequence length of 2,048.

    Left: the proportion of programming languages in CodeGeeX's training data. Right: the plot of training loss against the training steps of CodeGeeX.

    Code Corpus: Our training data contains two parts. The first part is from open-sourced code datasets, The Pile and CodeParrot. The Pile contains a subset of code corpus that collects public repositories with more than 100 stars from GitHub, from which we select codes in 23 popular programming languages. The second part is supplementary data directly scrapped from the public GitHub repositories that do not appear in previous datasets, including Python, Java and C++. To obtain data of potentially higher quality, repositories with at least one star and its size smaller than 10MB are chosen. A file is filtered out if it 1) has more than 100 characters per line on average, 2) is automatically generated, 3) has a ratio of alphabet less than 40%, or 4) is bigger than 100KB or smaller than 1KB. To help the model distinguish different languages, we add a language-specific prefix at the beginning of each segment in the form of [Comment sign] language: [LANG], e.g., # language: Python. For tokenization, we use the same tokenizer as GPT-2 and process whitespaces as extra tokens, resulting in a vocabulary of 50,400 tokens. In total, the code corpus has 23 programming languages with 158.7B tokens.

    Training: We implement CodeGeeX in Mindspore 1.7 and train it on 1,536 Ascend 910 AI Processor (32GB). The model weights are under FP16 format, except that we use FP32 for layer-norm and softmax for higher precision and stability. The entire model consumes about 27GB of memory. To increase the training efficiency, we adopt an 8-way model parallel training together with 192-way data parallel training, with ZeRO-2 optimizer enabled. The micro-batch size is 16 and the global batch size reaches 3,072. Moreover, we adopt techniques to further boost the training efficiency including the element-wise operator fusion, fast gelu activation, matrix multiplication dimension optimization, etc. The entire training process takes nearly two months, spanning from April 18 to June 22, 2022, during which 850B tokens were passed for training, i.e., 5+ epochs.

    To better evaluate the multilingual ability of code generation models, we propose a new benchmark HumanEval-X. While previous works evaluate multilingual program synthesis under semantic similarity (e.g., CodeBLEU) which is often misleading, HumanEval-X evaluates the functional correctness of the generated programs. HumanEval-X consists of 820 high-quality human-crafted data samples (each with test cases) in Python, C++, Java, JavaScript, and Go, and can be used for various tasks.

    An illustration of tasks supported by HumanEval-X. Declarations, docstrings, and solutions are marked with red, green, and blue respectively. Code generation uses declaration and docstring as input, to generate solution. Code translation uses declaration in both languages and translate the solution in source language to the one in target language.

    Our code is licensed under the Apache-2.0 license. Our model is licensed under the license.

    If you find our work useful, please cite:

  3. Jul 16, 2024 · CodeGeeX4, also known as CodeGeeX4-ALL-9B (part of same model series), is an open-source multilingual code generation model. It represents the latest in the CodeGeeX series and has been...

    • My Social
  4. Aug 4, 2023 · Large pre-trained code generation models, such as OpenAI Codex, can generate syntax-and function-correct code, making the coding of programmers more productive. In this paper, we introduce CodeGeeX, a multilingual model with 13 billion parameters for code generation.

  5. CodeGeeX can automatically generate code based on natural language descriptions of functions or generate subsequent code snippets based on existing code, enabling developers to quickly complete...

  6. We introduce CodeGeeX, a large-scale multilingual code generation model with 13 billion parameters, pre-trained on a large code corpus of more than 20 programming languages. As of June 22, 2022,...

  7. codegeex.cn › en-USCodeGeeX

    CodeGeeX is an AI-based coding assistant, which can suggest code in the current or following lines. It is powered by a large-scale multilingual code generation model with 13 billion parameters, pretrained on a large code corpus of more than 20 programming languages.

  1. Ad

    related to: Is codegeex a good AI code generator?
  2. Cody uses a deep understanding of your codebase to explain, fix, and help you write code. Cody helps you code faster with AI-assisted auto-complete in any programming language.

  1. People also search for