Skip to content
  • P
    Projects
  • G
    Groups
  • S
    Snippets
  • Help

SeetaResearch / Dragon

  • This project
    • Loading...
  • Sign in
Go to a project
  • Project
  • Repository
  • Issues 0
  • Merge Requests 0
  • Pipelines
  • Wiki
  • Snippets
  • Settings
  • Activity
  • Graph
  • Charts
  • Create a new issue
  • Jobs
  • Commits
  • Issue Boards
  • Files
  • Commits
  • Branches
  • Tags
  • Contributors
  • Graph
  • Compare
  • Charts
Switch branch/tag
  • dragon
  • torch
  • core
  • nn
  • functional.py
  • Ting PAN's avatar
    Use FP32 accumulator for FP16 ReduceSum · d56e67d1
    Summary:
    This commit adds a fallback with FP32 accumulator
    for FP16 ReduceSum to avoid dropping too many small values.
    Besides, FP16 kernels for arch < 530 are almost available.
    Ting PAN committed Nov 05, 2020
    d56e67d1
functional.py 45.6 KB
BlameHistoryPermalink