Skip to content
GitLab
Explore
Sign in
Primary navigation
Search or go to…
Project
H
hpvm-release
Manage
Activity
Members
Labels
Plan
Issues
Issue boards
Milestones
Wiki
Code
Merge requests
Repository
Branches
Commits
Tags
Repository graph
Compare revisions
Snippets
Build
Pipelines
Jobs
Pipeline schedules
Artifacts
Deploy
Releases
Model registry
Operate
Environments
Monitor
Incidents
Analyze
Value stream analytics
Contributor analytics
CI/CD analytics
Repository analytics
Model experiments
Help
Help
Support
GitLab documentation
Compare GitLab plans
Community forum
Contribute to GitLab
Provide feedback
Keyboard shortcuts
?
Snippets
Groups
Projects
Show more breadcrumbs
llvm
hpvm-release
Commits
b66caa18
Commit
b66caa18
authored
9 years ago
by
Prakalp Srivastava
Browse files
Options
Downloads
Patches
Plain Diff
Added old text to Related work
parent
68546cef
No related branches found
Branches containing commit
No related tags found
Tags containing commit
No related merge requests found
Changes
3
Hide whitespace changes
Inline
Side-by-side
Showing
3 changed files
paper/Abstract.tex
+2
-2
2 additions, 2 deletions
paper/Abstract.tex
paper/Compilation.tex
+1
-1
1 addition, 1 deletion
paper/Compilation.tex
paper/RelatedWork.tex
+22
-3
22 additions, 3 deletions
paper/RelatedWork.tex
with
25 additions
and
6 deletions
paper/Abstract.tex
+
2
−
2
View file @
b66caa18
...
...
@@ -7,14 +7,14 @@ However, programming such systems can be extremely complex as a single
SoC combines multiple different
parallelism models, instruction sets, and memory hierarchies, and different
SoCs use
\emph
{
different combinations
}
of these features.
We propose a new Virtual Instruction Set Architecture (ISA) that aims to
We propose
\NAME
{}
,
a new Virtual Instruction Set Architecture (ISA) that aims to
address both functional portability and performance portability across
mobile heterogeneous SoCs by capturing the wide range of different
parallelism models expected to be available on future SoCs.
Our virtual ISA design uses only two parallelism models to achieve this goal:
\emph
{
a hierarchical dataflow graph with side effects
}
and
\emph
{
parametric vector instructions
}
.
Our virtual ISA
is more general than existing ones that focus heavily on GPUs,
\NAME
{}
is more general than existing ones that focus heavily on GPUs,
such as PTX, HSAIL and SPIR, e.g., it can capture both streaming pipelined
parallelism and general dataflow parallelism found in many custom and
semi-custom (programmable) accelerators.
...
...
This diff is collapsed.
Click to expand it.
paper/Compilation.tex
+
1
−
1
View file @
b66caa18
...
...
@@ -98,7 +98,7 @@ Algorithm~\ref{algo:traversal} to traverse the graph of any internal node $N$.
\State
$
L
\gets
$
list of all nodes of
$
G
$
in topological order
\While
{$
L
$
is non-empty
}
\State
remove a node
$
n
$
from
$
L
$
\State
$
\textsc
{
CodeGen
}
(
n
)
$
\State
$
\textsc
{
Visit
}
(
n
)
$
\EndWhile
\EndIf
\EndProcedure
...
...
This diff is collapsed.
Click to expand it.
paper/RelatedWork.tex
+
22
−
3
View file @
b66caa18
\section
{
Related Work
}
\label
{
sec:related
}
\todo
[inline]
{
Source level programming languages
}
\todo
[inline]
{
Object code representation
}
\todo
[inline]
{
Compiler techniques - Tangram, Samar
}
\todo
[inline]
{
Source level programming languages OpenCL, CUDA, Renderscript,
Delite, ispc, halide
}
There are currently a few projects aiming to address the programmability and
portability challenges of SoC applications, but the ones focused at the source
level such as OpenCL~
\cite
{
OpenCL
}
, Lime~
\ref
{
Lime
}
and
Renderscript~
\cite
{
Renderscript
}
, are inadequate because they usually do not
provide adequately general parallelism models, which makes them unsuitable for
general purpose computation, and none of them provides object code portability,
with the exception of Renderscript. Renderscript, however, is tailored for image
filtering applications and uses LLVM scalar and vector code as its portable
object code format. Thus, it does not provide appropriate abstractions for the
different classes of parallel hardware.
\todo
[inline]
{
Object code representation HSAIL, SPIR, PTX
}
The ones focused on the object level
such as nVidia’s PTX~
\cite
{
PTX2.3Manual:URL
}
, OpenCL SPIR~
\cite
{
SPIRKhronosSpec
}
, HSAIL~
\cite
{
HSAIL
}
are either limited in scope of
supported hardware (for instance, PTX is designed specifically for nVidia
GPUs) or expose a restrictive throughput-oriented data parallelism and no other
models like dataflow-style or task parallelsim required across different
hardware.
\todo
[inline]
{
Compiler techniques - Tangram, Samar Amarasinghe's work
}
This diff is collapsed.
Click to expand it.
Preview
0%
Loading
Try again
or
attach a new file
.
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Save comment
Cancel
Please
register
or
sign in
to comment