MatsuLab. Lecture Note

ハイパフォーマンスコンピューティング

日時
月曜日 10:45〜12:15(3〜4限)
場所
西8号館 832号室
連絡
松岡教授 (Prof. S.Matsuoka)matsu あっと is.
TA 金刺 (H.Kanezashi)kanezashi あっと matsulab.is.
メーリングリストに追加しますので、至急TAまでメールを送ってください。Please email to Kanezashi (TA) as soon as possible in order to add you to the mailing list.

目次

休講予定日 Lecture Cancelled

10/19, 11/16, 02/08(補講はありません We do not have supplementary lectures)

授業概要と参考資料 Guidance and References

発表スケジュール Schedule

暫定的な割り当ては以下の通りですが、都合が悪い場合はTAまで希望日をメールしてください。

禁止リスト Inhibited List

  • Training Large Scale Deep Neural Networks on the Intel Xeon Phi Many-Core Coprocessor
  • Memory fast-forward: A low cost special function unit to enhance energy efficiency in GPU for big data processing
  • Optimized Deep Learning Architectures with Fast Matrix Operation Kernels on Parallel Platform
  • Real-time anomaly detection in hyperspectral images using multivariate normal mixture models and GPU processing
  • Hyperspectral Unmixing on GPUs and Multi-Core Processors: A Comparison
  • Performance versus energy consumption of hyperspectral unmixing algorithms on multi-core platforms
  • Optimizing communication and cooling costs in HPC data centers via intelligent job allocation
  • Cost Minimization for Big Data Processing in Geo-Distributed Data Centers
  • On Characterization of Performance and Energy Efficiency in Heterogeneous HPC Cloud Data Centers
  • DaDianNao?: A Machine-Learning Supercomputer
  • Mariana: tencent deep learning platform and its applications
  • Performance Modeling and Scalability Optimization of Distributed Deep Learning Systems
  • Asynchronous parallel stochastic gradient descent: a numeric core for scalable distributed machine learning algorithms
  • FireCaffe?: near-linear acceleration of deep neural network training on compute clusters
  • CA-SVM: Communication-Avoiding Support Vector Machines on Distributed Systems
  • Large Scale Distributed Deep Networks
  • Optimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks
  • 24.77 Pflops on a Gravitational Tree-Code to Simulate the Milky Way Galaxy with 18600 GPUs
  • Massively Parallel Models of the Human Circulatory System
  • Moving to memoryland: in-memory computation for existing applications
  • Intelligent SSD: A Turbo for Big Data Mining
  • Scalable Multi-Access Flash Store for Big Data Analytics
  • An FPGA-Based Tightly Coupled Accelerator for Data-Intensive Applications
  • A reconfigurable fabric for accelerating large-scale datacenter services
  • An FPGA-based In-Line Accelerator for Memcached
  • Scaling Up the Training of Deep CNNs for Human Action Recognition

期末レポート Report

  • 期限 Due date: 02/17 (Extended)
  • Summarize the general topic covering and including ALL THREE PAPERS regarding the state of the art in HPC and Big Data convergence.
  • It should be 10 pages in IEEE conference paper format
  • Please submit it to TA by email (NOT mailing list)

リンク Links


添付ファイル: file06853195.pdf 1817件 [詳細] fileHPC_Class_Presentation_Hamid.pdf 2006件 [詳細] filefpga2014-wjun.pdf 1754件 [詳細] filep161-zhang-small.pdf 67597件 [詳細] filehpc_Kanezashi1.pdf 1735件 [詳細] fileOptimizing FPGA-based Accelerator Design for Deep Convolutional Neural Networks.pdf 8635件 [詳細] fileHPC15_2015-12-14_Jian_FireCaffe.pdf 2026件 [詳細] fileFireCaffe.pdf 1747件 [詳細] filehpc_teranishi.pdf 1295件 [詳細] filea1-keuper.pdf 1470件 [詳細] fileDianNao.pdf 5800件 [詳細] fileDaDianNao.pdf 2745件 [詳細] filehpc_Motoyama.pdf 1397件 [詳細] fileHyperspectral.pdf 1858件 [詳細] fileHPC15_2015-12-14&21_Jian_FireCaffe v2.pdf 1069件 [詳細] fileHPC_kuroda.pdf 1589件 [詳細] filehpc_uehara.pdf 1732件 [詳細] fileHPC_Motoyama2.pdf 1622件 [詳細] filehpc_teranishi2.pdf 1363件 [詳細] file2015年度ハイパフォーマンスコンピューティング授業内容.docx 397件 [詳細] file06735232.pdf 1459件 [詳細] file2015年度ハイパフォーマンスコンピューティング授業内容.pdf 1962件 [詳細]

トップ   編集 凍結 差分 バックアップ 添付 複製 名前変更 リロード   新規 一覧 単語検索 最終更新   ヘルプ   最終更新のRSS
Last-modified: 2018-05-29 (火) 19:12:57 (2151d)