Skip to content

Commit e52c19f

Browse files
authored
[DOCS] Update README (#17650)
* [DOCS] Correct the "Getting Started" link
1 parent 73f6c49 commit e52c19f

File tree

1 file changed

+13
-13
lines changed

1 file changed

+13
-13
lines changed

README.md

Lines changed: 13 additions & 13 deletions
Original file line numberDiff line numberDiff line change
@@ -23,8 +23,8 @@
2323
[Release Notes](NEWS.md)
2424

2525
Apache TVM is a compiler stack for deep learning systems. It is designed to close the gap between the
26-
productivity-focused deep learning frameworks, and the performance- and efficiency-focused hardware backends.
27-
TVM works with deep learning frameworks to provide end to end compilation to different backends.
26+
productivity-focused deep learning frameworks and the performance- and efficiency-focused hardware backends.
27+
TVM works with deep learning frameworks to provide end-to-end compilation for different backends.
2828

2929
License
3030
-------
@@ -33,31 +33,31 @@ TVM is licensed under the [Apache-2.0](LICENSE) license.
3333
Getting Started
3434
---------------
3535
Check out the [TVM Documentation](https://tvm.apache.org/docs/) site for installation instructions, tutorials, examples, and more.
36-
The [Getting Started with TVM](https://tvm.apache.org/docs/tutorial/introduction.html) tutorial is a great
36+
The [Getting Started with TVM](https://tvm.apache.org/docs/get_started/overview.html) tutorial is a great
3737
place to start.
3838

3939
Contribute to TVM
4040
-----------------
41-
TVM adopts apache committer model, we aim to create an open source project that is maintained and owned by the community.
41+
TVM adopts the Apache committer model. We aim to create an open-source project maintained and owned by the community.
4242
Check out the [Contributor Guide](https://tvm.apache.org/docs/contribute/).
4343

4444
History and Acknowledgement
4545
---------------------------
46-
TVM started as a research project for deep learning compiler.
47-
The first version of the project benefited a lot from following projects:
46+
TVM started as a research project for deep learning compilation.
47+
The first version of the project benefited a lot from the following projects:
4848

4949
- [Halide](https://github.com/halide/Halide): Part of TVM's TIR and arithmetic simplification module
50-
originates from Halide. We also learned and adapted some part of lowering pipeline from Halide.
50+
originates from Halide. We also learned and adapted some parts of the lowering pipeline from Halide.
5151
- [Loopy](https://github.com/inducer/loopy): use of integer set analysis and its loop transformation primitives.
5252
- [Theano](https://github.com/Theano/Theano): the design inspiration of symbolic scan operator for recurrence.
5353

5454
Since then, the project has gone through several rounds of redesigns.
5555
The current design is also drastically different from the initial design, following the
56-
development trend of ML compiler community.
56+
development trend of the ML compiler community.
5757

58-
The most recent version focuses on a cross-level design with TensorIR as tensor-level representation
59-
and Relax as graph level representation, and python-first transformations.
60-
The current design goal of the project is to make the ML compiler accessible by enabling most
58+
The most recent version focuses on a cross-level design with TensorIR as the tensor-level representation
59+
and Relax as the graph-level representation and Python-first transformations.
60+
The project's current design goal is to make the ML compiler accessible by enabling most
6161
transformations to be customizable in Python and bringing a cross-level representation that can jointly
62-
optimize computational graphs, tensor programs, and libraries. The project also serves as a foundation
63-
infra to build python-first vertical compilers for various domains, such as LLMs.
62+
optimize computational graphs, tensor programs, and libraries. The project is also a foundation
63+
infra for building Python-first vertical compilers for domains, such as LLMs.

0 commit comments

Comments
 (0)