book polish

This commit is contained in:
Per Lindgren 2023-02-01 19:46:58 +01:00 committed by Henrik Tjäder
parent 14fdca130f
commit 89632f9b22
11 changed files with 90 additions and 26 deletions

View file

@ -17,6 +17,9 @@ The example below shows that `idle` runs after `init`.
``` console
$ cargo run --target thumbv7m-none-eabi --example idle
```
``` console
{{#include ../../../../rtic/ci/expected/idle.run}}
```
@ -41,6 +44,9 @@ The following example shows how to enable sleep by setting the
``` console
$ cargo run --target thumbv7m-none-eabi --example idle-wfi
```
``` console
{{#include ../../../../rtic/ci/expected/idle-wfi.run}}
```

View file

@ -24,5 +24,8 @@ Running the example will print `init` to the console and then exit the QEMU proc
``` console
$ cargo run --target thumbv7m-none-eabi --example init
```
``` console
{{#include ../../../../rtic/ci/expected/init.run}}
```

View file

@ -11,6 +11,9 @@ RTIC is designed with resource efficiency in mind. RTIC itself does not rely on
For a minimal example you can expect something like:
``` console
$ cargo size --example smallest --target thumbv7m-none-eabi --release
```
``` console
Finished release [optimized] target(s) in 0.07s
text data bss dec hex filename
924 0 0 924 39c smallest

View file

@ -2,7 +2,7 @@
Channels can be used to communicate data between running *software* tasks. The channel is essentially a wait queue, allowing tasks with multiple producers and a single receiver. A channel is constructed in the `init` task and backed by statically allocated memory. Send and receive endpoints are distributed to *software* tasks:
```rust
``` rust
...
const CAPACITY: usize = 5;
#[init]
@ -20,7 +20,7 @@ In this case the channel holds data of `u32` type with a capacity of 5 elements
The `send` method post a message on the channel as shown below:
```rust
``` rust
#[task]
async fn sender1(_c: sender1::Context, mut sender: Sender<'static, u32, CAPACITY>) {
hprintln!("Sender 1 sending: 1");
@ -32,7 +32,7 @@ async fn sender1(_c: sender1::Context, mut sender: Sender<'static, u32, CAPACITY
The receiver can `await` incoming messages:
```rust
``` rust
#[task]
async fn receiver(_c: receiver::Context, mut receiver: Receiver<'static, u32, CAPACITY>) {
while let Ok(val) = receiver.recv().await {
@ -42,6 +42,8 @@ async fn receiver(_c: receiver::Context, mut receiver: Receiver<'static, u32, CA
}
```
Channels are implemented using a small (global) *Critical Section* (CS) for protection against race-conditions. The user must provide an CS implementation. Compiling the examples given the `--features test-critical-section` gives one possible implementation.
For a complete example:
``` rust
@ -50,6 +52,9 @@ For a complete example:
``` console
$ cargo run --target thumbv7m-none-eabi --example async-channel --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-channel.run}}
```
@ -79,7 +84,10 @@ In case all senders have been dropped `await` on an empty receiver channel resul
```
``` console
$ cargo run --target thumbv7m-none-eabi --example async-channel-no-sender --features test-critical-section
$ cargo run --target thumbv7m-none-eabi --example async-channel-no-sender --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-channel-no-sender.run}}
```
@ -93,6 +101,9 @@ The resulting error returns the data back to the sender, allowing the sender to
``` console
$ cargo run --target thumbv7m-none-eabi --example async-channel-no-receiver --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-channel-no-receiver.run}}
```
@ -107,6 +118,9 @@ In cases you wish the sender to proceed even in case the channel is full. To tha
```
``` console
$ cargo run --target thumbv7m-none-eabi --example async-channel-try --features test-critical-section
$ cargo run --target thumbv7m-none-eabi --example async-channel-try --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-channel-try.run}}
```

View file

@ -4,7 +4,7 @@ A convenient way to express *miniminal* timing requirements is by means of delay
This can be achieved by instantiating a monotonic timer:
```rust
``` rust
...
rtic_monotonics::make_systick_timer_queue!(TIMER);
@ -17,7 +17,7 @@ fn init(cx: init::Context) -> (Shared, Local) {
A *software* task can `await` the delay to expire:
```rust
``` rust
#[task]
async fn foo(_cx: foo::Context) {
...
@ -27,6 +27,8 @@ async fn foo(_cx: foo::Context) {
Technically, the timer queue is implemented as a list based priority queue, where list-nodes are statically allocated as part of the underlying task `Future`. Thus, the timer queue is infallible at run-time (its size and allocation is determined at compile time).
Similarly the channels implementation, the timer-queue implementation relies on a global *Critical Section* (CS) for race protection. For the examples a CS implementation is provided by adding `--features test-critical-section` to the build options.
For a complete example:
``` rust
@ -35,6 +37,9 @@ For a complete example:
``` console
$ cargo run --target thumbv7m-none-eabi --example async-delay --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-delay.run}}
```
@ -112,5 +117,8 @@ The complete example:
``` console
$ cargo run --target thumbv7m-none-eabi --example async-timeout --features test-critical-section
```
``` console
{{#include ../../../../rtic/ci/expected/async-timeout.run}}
```

View file

@ -25,5 +25,8 @@ The example below demonstrates the use of the `#[task(binds = InterruptName)]` a
``` console
$ cargo run --target thumbv7m-none-eabi --example hardware
```
``` console
{{#include ../../../../rtic/ci/expected/hardware.run}}
```

View file

@ -33,6 +33,9 @@ Running the example:
``` console
$ cargo run --target thumbv7m-none-eabi --example locals
```
``` console
{{#include ../../../../rtic/ci/expected/locals.run}}
```
@ -79,6 +82,9 @@ In the example below we have three interrupt handlers with priorities ranging fr
``` console
$ cargo run --target thumbv7m-none-eabi --example lock
```
``` console
{{#include ../../../../rtic/ci/expected/lock.run}}
```
@ -94,6 +100,9 @@ As an extension to `lock`, and to reduce rightward drift, locks can be taken as
``` console
$ cargo run --target thumbv7m-none-eabi --example multilock
```
``` console
{{#include ../../../../rtic/ci/expected/multilock.run}}
```
@ -113,6 +122,9 @@ In the example below a key (e.g. a cryptographic key) is loaded (or created) at
``` console
$ cargo run --target thumbv7m-none-eabi --example only-shared-access
```
``` console
{{#include ../../../../rtic/ci/expected/only-shared-access.run}}
```
@ -136,5 +148,8 @@ Using `#[lock_free]` on resources shared by tasks running at different prioritie
``` console
$ cargo run --target thumbv7m-none-eabi --example lock-free
```
``` console
{{#include ../../../../rtic/ci/expected/lock-free.run}}
```

View file

@ -29,6 +29,9 @@ See the following example:
``` console
$ cargo run --target thumbv7m-none-eabi --example spawn
```
``` console
{{#include ../../../../rtic/ci/expected/spawn.run}}
```
You may `spawn` a *software* task again, given that it has run-to-completion (returned).
@ -43,6 +46,9 @@ Technically the async executor will `poll` the `foo` *future* which in this case
``` console
$ cargo run --target thumbv7m-none-eabi --example spawn_loop
```
``` console
{{#include ../../../../rtic/ci/expected/spawn_loop.run}}
```
@ -56,6 +62,9 @@ Technically, a `spawn` to a *future* that is not in *completed* state is conside
``` console
$ cargo run --target thumbv7m-none-eabi --example spawn_err
```
``` console
{{#include ../../../../rtic/ci/expected/spawn_err.run}}
```
@ -68,6 +77,9 @@ You can also pass arguments at spawn as follows.
``` console
$ cargo run --target thumbv7m-none-eabi --example spawn_arguments
```
``` console
{{#include ../../../../rtic/ci/expected/spawn_arguments.run}}
```
@ -86,6 +98,9 @@ Conceptually, one can see such tasks as running in the `main` thread of the appl
``` console
$ cargo run --target thumbv7m-none-eabi --example zero-prio-task
```
``` console
{{#include ../../../../rtic/ci/expected/zero-prio-task.run}}
```

View file

@ -1,14 +1,15 @@
# Resource de-structure-ing
Destructuring task resources might help readability if a task takes multiple
resources.
Here are two examples on how to split up the resource struct:
resources. Here are two examples on how to split up the resource struct:
``` rust
{{#include ../../../../examples/destructure.rs}}
{{#include ../../../../rtic/examples/destructure.rs}}
```
``` console
$ cargo run --target thumbv7m-none-eabi --example destructure
{{#include ../../../../ci/expected/destructure.run}}
```
``` console
{{#include ../../../../rtic/ci/expected/destructure.run}}
```

View file

@ -1,31 +1,26 @@
# Using indirection for faster message passing
Message passing always involves copying the payload from the sender into a
static variable and then from the static variable into the receiver. Thus
sending a large buffer, like a `[u8; 128]`, as a message involves two expensive
Message passing always involves copying the payload from the sender into a static variable and then from the static variable into the receiver. Thus sending a large buffer, like a `[u8; 128]`, as a message involves two expensive
`memcpy`s.
Indirection can minimize message passing overhead:
instead of sending the buffer by value, one can send an owning pointer into the
buffer.
Indirection can minimize message passing overhead: instead of sending the buffer by value, one can send an owning pointer into the buffer.
One can use a global memory allocator to achieve indirection (`alloc::Box`,
`alloc::Rc`, etc.), which requires using the nightly channel as of Rust v1.37.0,
or one can use a statically allocated memory pool like [`heapless::Pool`].
One can use a global memory allocator to achieve indirection (`alloc::Box`, `alloc::Rc`, etc.), which requires using the nightly channel as of Rust v1.37.0, or one can use a statically allocated memory pool like [`heapless::Pool`].
[`heapless::Pool`]: https://docs.rs/heapless/0.5.0/heapless/pool/index.html
As this example of approach goes completely outside of RTIC resource
model with shared and local the program would rely on the correctness
of the memory allocator, in this case `heapless::pool`.
As this example of approach goes completely outside of RTIC resource model with shared and local the program would rely on the correctness of the memory allocator, in this case `heapless::pool`.
Here's an example where `heapless::Pool` is used to "box" buffers of 128 bytes.
``` rust
{{#include ../../../../examples/pool.rs}}
{{#include ../../../../rtic/examples/pool.rs}}
```
``` console
$ cargo run --target thumbv7m-none-eabi --example pool
{{#include ../../../../ci/expected/pool.run}}
```
``` console
{{#include ../../../../rtic/ci/expected/pool.run}}
```

View file

@ -4,6 +4,7 @@
#![deny(warnings)]
#![no_main]
#![no_std]
#![feature(type_alias_impl_trait)]
use heapless::{
pool,
@ -41,7 +42,7 @@ mod app {
}
#[task(binds = I2C0, priority = 2)]
async fn i2c0(_: i2c0::Context) {
fn i2c0(_: i2c0::Context) {
// claim a memory block, initialize it and ..
let x = P::alloc().unwrap().init([0u8; 128]);