diff --git a/book/en/src/by-example/app_idle.md b/book/en/src/by-example/app_idle.md index 4856ee19a6..cbfd7ba50b 100644 --- a/book/en/src/by-example/app_idle.md +++ b/book/en/src/by-example/app_idle.md @@ -17,6 +17,9 @@ The example below shows that `idle` runs after `init`. ``` console $ cargo run --target thumbv7m-none-eabi --example idle +``` + +``` console {{#include ../../../../rtic/ci/expected/idle.run}} ``` @@ -41,6 +44,9 @@ The following example shows how to enable sleep by setting the ``` console $ cargo run --target thumbv7m-none-eabi --example idle-wfi +``` + +``` console {{#include ../../../../rtic/ci/expected/idle-wfi.run}} ``` diff --git a/book/en/src/by-example/app_init.md b/book/en/src/by-example/app_init.md index 3767bd7616..fb37387ba2 100644 --- a/book/en/src/by-example/app_init.md +++ b/book/en/src/by-example/app_init.md @@ -24,5 +24,8 @@ Running the example will print `init` to the console and then exit the QEMU proc ``` console $ cargo run --target thumbv7m-none-eabi --example init +``` + +``` console {{#include ../../../../rtic/ci/expected/init.run}} ``` diff --git a/book/en/src/by-example/app_minimal.md b/book/en/src/by-example/app_minimal.md index f241089061..714f543229 100644 --- a/book/en/src/by-example/app_minimal.md +++ b/book/en/src/by-example/app_minimal.md @@ -11,6 +11,9 @@ RTIC is designed with resource efficiency in mind. RTIC itself does not rely on For a minimal example you can expect something like: ``` console $ cargo size --example smallest --target thumbv7m-none-eabi --release +``` + +``` console Finished release [optimized] target(s) in 0.07s text data bss dec hex filename 924 0 0 924 39c smallest diff --git a/book/en/src/by-example/channel.md b/book/en/src/by-example/channel.md index 99bfedda3c..1f9510aa22 100644 --- a/book/en/src/by-example/channel.md +++ b/book/en/src/by-example/channel.md @@ -2,7 +2,7 @@ Channels can be used to communicate data between running *software* tasks. The channel is essentially a wait queue, allowing tasks with multiple producers and a single receiver. A channel is constructed in the `init` task and backed by statically allocated memory. Send and receive endpoints are distributed to *software* tasks: -```rust +``` rust ... const CAPACITY: usize = 5; #[init] @@ -20,7 +20,7 @@ In this case the channel holds data of `u32` type with a capacity of 5 elements The `send` method post a message on the channel as shown below: -```rust +``` rust #[task] async fn sender1(_c: sender1::Context, mut sender: Sender<'static, u32, CAPACITY>) { hprintln!("Sender 1 sending: 1"); @@ -32,7 +32,7 @@ async fn sender1(_c: sender1::Context, mut sender: Sender<'static, u32, CAPACITY The receiver can `await` incoming messages: -```rust +``` rust #[task] async fn receiver(_c: receiver::Context, mut receiver: Receiver<'static, u32, CAPACITY>) { while let Ok(val) = receiver.recv().await { @@ -42,6 +42,8 @@ async fn receiver(_c: receiver::Context, mut receiver: Receiver<'static, u32, CA } ``` +Channels are implemented using a small (global) *Critical Section* (CS) for protection against race-conditions. The user must provide an CS implementation. Compiling the examples given the `--features test-critical-section` gives one possible implementation. + For a complete example: ``` rust @@ -50,6 +52,9 @@ For a complete example: ``` console $ cargo run --target thumbv7m-none-eabi --example async-channel --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-channel.run}} ``` @@ -79,7 +84,10 @@ In case all senders have been dropped `await` on an empty receiver channel resul ``` ``` console -$ cargo run --target thumbv7m-none-eabi --example async-channel-no-sender --features test-critical-section +$ cargo run --target thumbv7m-none-eabi --example async-channel-no-sender --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-channel-no-sender.run}} ``` @@ -93,6 +101,9 @@ The resulting error returns the data back to the sender, allowing the sender to ``` console $ cargo run --target thumbv7m-none-eabi --example async-channel-no-receiver --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-channel-no-receiver.run}} ``` @@ -107,6 +118,9 @@ In cases you wish the sender to proceed even in case the channel is full. To tha ``` ``` console -$ cargo run --target thumbv7m-none-eabi --example async-channel-try --features test-critical-section +$ cargo run --target thumbv7m-none-eabi --example async-channel-try --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-channel-try.run}} ``` \ No newline at end of file diff --git a/book/en/src/by-example/delay.md b/book/en/src/by-example/delay.md index d35d9da453..8d05d7ce01 100644 --- a/book/en/src/by-example/delay.md +++ b/book/en/src/by-example/delay.md @@ -4,7 +4,7 @@ A convenient way to express *miniminal* timing requirements is by means of delay This can be achieved by instantiating a monotonic timer: -```rust +``` rust ... rtic_monotonics::make_systick_timer_queue!(TIMER); @@ -17,7 +17,7 @@ fn init(cx: init::Context) -> (Shared, Local) { A *software* task can `await` the delay to expire: -```rust +``` rust #[task] async fn foo(_cx: foo::Context) { ... @@ -27,6 +27,8 @@ async fn foo(_cx: foo::Context) { Technically, the timer queue is implemented as a list based priority queue, where list-nodes are statically allocated as part of the underlying task `Future`. Thus, the timer queue is infallible at run-time (its size and allocation is determined at compile time). +Similarly the channels implementation, the timer-queue implementation relies on a global *Critical Section* (CS) for race protection. For the examples a CS implementation is provided by adding `--features test-critical-section` to the build options. + For a complete example: ``` rust @@ -35,6 +37,9 @@ For a complete example: ``` console $ cargo run --target thumbv7m-none-eabi --example async-delay --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-delay.run}} ``` @@ -112,5 +117,8 @@ The complete example: ``` console $ cargo run --target thumbv7m-none-eabi --example async-timeout --features test-critical-section +``` + +``` console {{#include ../../../../rtic/ci/expected/async-timeout.run}} ``` diff --git a/book/en/src/by-example/hardware_tasks.md b/book/en/src/by-example/hardware_tasks.md index cb20a7ccd4..c9022674bd 100644 --- a/book/en/src/by-example/hardware_tasks.md +++ b/book/en/src/by-example/hardware_tasks.md @@ -25,5 +25,8 @@ The example below demonstrates the use of the `#[task(binds = InterruptName)]` a ``` console $ cargo run --target thumbv7m-none-eabi --example hardware +``` + +``` console {{#include ../../../../rtic/ci/expected/hardware.run}} ``` diff --git a/book/en/src/by-example/resources.md b/book/en/src/by-example/resources.md index 2dd7cb7a4b..0bf5d11fb8 100644 --- a/book/en/src/by-example/resources.md +++ b/book/en/src/by-example/resources.md @@ -33,6 +33,9 @@ Running the example: ``` console $ cargo run --target thumbv7m-none-eabi --example locals +``` + +``` console {{#include ../../../../rtic/ci/expected/locals.run}} ``` @@ -79,6 +82,9 @@ In the example below we have three interrupt handlers with priorities ranging fr ``` console $ cargo run --target thumbv7m-none-eabi --example lock +``` + +``` console {{#include ../../../../rtic/ci/expected/lock.run}} ``` @@ -94,6 +100,9 @@ As an extension to `lock`, and to reduce rightward drift, locks can be taken as ``` console $ cargo run --target thumbv7m-none-eabi --example multilock +``` + +``` console {{#include ../../../../rtic/ci/expected/multilock.run}} ``` @@ -113,6 +122,9 @@ In the example below a key (e.g. a cryptographic key) is loaded (or created) at ``` console $ cargo run --target thumbv7m-none-eabi --example only-shared-access +``` + +``` console {{#include ../../../../rtic/ci/expected/only-shared-access.run}} ``` @@ -136,5 +148,8 @@ Using `#[lock_free]` on resources shared by tasks running at different prioritie ``` console $ cargo run --target thumbv7m-none-eabi --example lock-free +``` + +``` console {{#include ../../../../rtic/ci/expected/lock-free.run}} ``` diff --git a/book/en/src/by-example/software_tasks.md b/book/en/src/by-example/software_tasks.md index 828c3fd8ce..0efc57ba02 100644 --- a/book/en/src/by-example/software_tasks.md +++ b/book/en/src/by-example/software_tasks.md @@ -29,6 +29,9 @@ See the following example: ``` console $ cargo run --target thumbv7m-none-eabi --example spawn +``` + +``` console {{#include ../../../../rtic/ci/expected/spawn.run}} ``` You may `spawn` a *software* task again, given that it has run-to-completion (returned). @@ -43,6 +46,9 @@ Technically the async executor will `poll` the `foo` *future* which in this case ``` console $ cargo run --target thumbv7m-none-eabi --example spawn_loop +``` + +``` console {{#include ../../../../rtic/ci/expected/spawn_loop.run}} ``` @@ -56,6 +62,9 @@ Technically, a `spawn` to a *future* that is not in *completed* state is conside ``` console $ cargo run --target thumbv7m-none-eabi --example spawn_err +``` + +``` console {{#include ../../../../rtic/ci/expected/spawn_err.run}} ``` @@ -68,6 +77,9 @@ You can also pass arguments at spawn as follows. ``` console $ cargo run --target thumbv7m-none-eabi --example spawn_arguments +``` + +``` console {{#include ../../../../rtic/ci/expected/spawn_arguments.run}} ``` @@ -86,6 +98,9 @@ Conceptually, one can see such tasks as running in the `main` thread of the appl ``` console $ cargo run --target thumbv7m-none-eabi --example zero-prio-task +``` + +``` console {{#include ../../../../rtic/ci/expected/zero-prio-task.run}} ``` diff --git a/book/en/src/by-example/tips_destructureing.md b/book/en/src/by-example/tips_destructureing.md index 4637b48343..ab27987a56 100644 --- a/book/en/src/by-example/tips_destructureing.md +++ b/book/en/src/by-example/tips_destructureing.md @@ -1,14 +1,15 @@ # Resource de-structure-ing Destructuring task resources might help readability if a task takes multiple -resources. -Here are two examples on how to split up the resource struct: +resources. Here are two examples on how to split up the resource struct: ``` rust -{{#include ../../../../examples/destructure.rs}} +{{#include ../../../../rtic/examples/destructure.rs}} ``` ``` console $ cargo run --target thumbv7m-none-eabi --example destructure -{{#include ../../../../ci/expected/destructure.run}} +``` +``` console +{{#include ../../../../rtic/ci/expected/destructure.run}} ``` diff --git a/book/en/src/by-example/tips_indirection.md b/book/en/src/by-example/tips_indirection.md index 567a5e723b..0de14a612d 100644 --- a/book/en/src/by-example/tips_indirection.md +++ b/book/en/src/by-example/tips_indirection.md @@ -1,31 +1,26 @@ # Using indirection for faster message passing -Message passing always involves copying the payload from the sender into a -static variable and then from the static variable into the receiver. Thus -sending a large buffer, like a `[u8; 128]`, as a message involves two expensive +Message passing always involves copying the payload from the sender into a static variable and then from the static variable into the receiver. Thus sending a large buffer, like a `[u8; 128]`, as a message involves two expensive `memcpy`s. -Indirection can minimize message passing overhead: -instead of sending the buffer by value, one can send an owning pointer into the -buffer. +Indirection can minimize message passing overhead: instead of sending the buffer by value, one can send an owning pointer into the buffer. -One can use a global memory allocator to achieve indirection (`alloc::Box`, -`alloc::Rc`, etc.), which requires using the nightly channel as of Rust v1.37.0, -or one can use a statically allocated memory pool like [`heapless::Pool`]. +One can use a global memory allocator to achieve indirection (`alloc::Box`, `alloc::Rc`, etc.), which requires using the nightly channel as of Rust v1.37.0, or one can use a statically allocated memory pool like [`heapless::Pool`]. [`heapless::Pool`]: https://docs.rs/heapless/0.5.0/heapless/pool/index.html -As this example of approach goes completely outside of RTIC resource -model with shared and local the program would rely on the correctness -of the memory allocator, in this case `heapless::pool`. +As this example of approach goes completely outside of RTIC resource model with shared and local the program would rely on the correctness of the memory allocator, in this case `heapless::pool`. Here's an example where `heapless::Pool` is used to "box" buffers of 128 bytes. ``` rust -{{#include ../../../../examples/pool.rs}} +{{#include ../../../../rtic/examples/pool.rs}} ``` ``` console $ cargo run --target thumbv7m-none-eabi --example pool -{{#include ../../../../ci/expected/pool.run}} +``` + +``` console +{{#include ../../../../rtic/ci/expected/pool.run}} ``` diff --git a/rtic/examples/pool.no_rs b/rtic/examples/pool.rs similarity index 95% rename from rtic/examples/pool.no_rs rename to rtic/examples/pool.rs index fb8589ad46..e597ec69f5 100644 --- a/rtic/examples/pool.no_rs +++ b/rtic/examples/pool.rs @@ -4,6 +4,7 @@ #![deny(warnings)] #![no_main] #![no_std] +#![feature(type_alias_impl_trait)] use heapless::{ pool, @@ -41,7 +42,7 @@ mod app { } #[task(binds = I2C0, priority = 2)] - async fn i2c0(_: i2c0::Context) { + fn i2c0(_: i2c0::Context) { // claim a memory block, initialize it and .. let x = P::alloc().unwrap().init([0u8; 128]);