|
1 | 1 | /* automatically generated by rust-bindgen */
|
2 | 2 |
|
3 |
| -#![allow(dead_code, non_snake_case, non_camel_case_types, non_upper_case_globals)] |
| 3 | +#![allow( |
| 4 | + dead_code, |
| 5 | + non_snake_case, |
| 6 | + non_camel_case_types, |
| 7 | + non_upper_case_globals |
| 8 | +)] |
4 | 9 |
|
5 | 10 | pub const RTE_CACHE_LINE_SIZE: u32 = 64;
|
6 | 11 | pub const RTE_MEMPOOL_OPS_NAMESIZE: u32 = 32;
|
@@ -46,17 +51,17 @@ pub type rte_mempool_get_count =
|
46 | 51 | #[repr(C)]
|
47 | 52 | #[derive(Copy, Clone)]
|
48 | 53 | pub struct rte_mempool_ops {
|
49 |
| - /// < Name of mempool ops struct. |
| 54 | + ///< Name of mempool ops struct. |
50 | 55 | pub name: [::std::os::raw::c_char; 32usize],
|
51 |
| - /// < Allocate private data. |
| 56 | + ///< Allocate private data. |
52 | 57 | pub alloc: rte_mempool_alloc_t,
|
53 |
| - /// < Free the external pool. |
| 58 | + ///< Free the external pool. |
54 | 59 | pub free: rte_mempool_free_t,
|
55 |
| - /// < Enqueue an object. |
| 60 | + ///< Enqueue an object. |
56 | 61 | pub enqueue: rte_mempool_enqueue_t,
|
57 |
| - /// < Dequeue an object. |
| 62 | + ///< Dequeue an object. |
58 | 63 | pub dequeue: rte_mempool_dequeue_t,
|
59 |
| - /// < Get qty of available objs. |
| 64 | + ///< Get qty of available objs. |
60 | 65 | pub get_count: rte_mempool_get_count,
|
61 | 66 | pub __bindgen_padding_0: [u64; 7usize],
|
62 | 67 | }
|
@@ -135,16 +140,19 @@ impl Default for rte_mempool_ops {
|
135 | 140 | }
|
136 | 141 | impl ::std::cmp::PartialEq for rte_mempool_ops {
|
137 | 142 | fn eq(&self, other: &rte_mempool_ops) -> bool {
|
138 |
| - self.name == other.name && self.alloc == other.alloc && self.free == other.free |
139 |
| - && self.enqueue == other.enqueue && self.dequeue == other.dequeue |
| 143 | + self.name == other.name |
| 144 | + && self.alloc == other.alloc |
| 145 | + && self.free == other.free |
| 146 | + && self.enqueue == other.enqueue |
| 147 | + && self.dequeue == other.dequeue |
140 | 148 | && self.get_count == other.get_count
|
141 | 149 | }
|
142 | 150 | }
|
143 | 151 | /// The rte_spinlock_t type.
|
144 | 152 | #[repr(C)]
|
145 | 153 | #[derive(Debug, Default, Copy, Clone, Hash, PartialEq, Eq)]
|
146 | 154 | pub struct rte_spinlock_t {
|
147 |
| - /// < lock status 0 = unlocked, 1 = locked |
| 155 | + ///< lock status 0 = unlocked, 1 = locked |
148 | 156 | pub locked: ::std::os::raw::c_int,
|
149 | 157 | }
|
150 | 158 | #[test]
|
@@ -180,9 +188,9 @@ fn bindgen_test_layout_rte_spinlock_t() {
|
180 | 188 | #[repr(C)]
|
181 | 189 | #[derive(Copy, Clone)]
|
182 | 190 | pub struct rte_mempool_ops_table {
|
183 |
| - /// < Spinlock for add/delete. |
| 191 | + ///< Spinlock for add/delete. |
184 | 192 | pub sl: rte_spinlock_t,
|
185 |
| - /// < Number of used ops structs in the table. |
| 193 | + ///< Number of used ops structs in the table. |
186 | 194 | pub num_ops: u32,
|
187 | 195 | pub __bindgen_padding_0: [u64; 7usize],
|
188 | 196 | /// Storage for all possible ops structs.
|
@@ -330,8 +338,10 @@ impl Default for malloc_heap {
|
330 | 338 | }
|
331 | 339 | impl ::std::cmp::PartialEq for malloc_heap {
|
332 | 340 | fn eq(&self, other: &malloc_heap) -> bool {
|
333 |
| - self.lock == other.lock && self.free_head == other.free_head |
334 |
| - && self.alloc_count == other.alloc_count && self.total_size == other.total_size |
| 341 | + self.lock == other.lock |
| 342 | + && self.free_head == other.free_head |
| 343 | + && self.alloc_count == other.alloc_count |
| 344 | + && self.total_size == other.total_size |
335 | 345 | }
|
336 | 346 | }
|
337 | 347 | #[repr(C)]
|
|
0 commit comments