Commit | Line | Data |
---|---|---|
77c9a5da SR |
1 | #ifndef _LINUX_FIREWIRE_H |
2 | #define _LINUX_FIREWIRE_H | |
3 | ||
4 | #include <linux/completion.h> | |
5 | #include <linux/device.h> | |
c76acec6 | 6 | #include <linux/dma-mapping.h> |
77c9a5da SR |
7 | #include <linux/kernel.h> |
8 | #include <linux/kref.h> | |
9 | #include <linux/list.h> | |
10 | #include <linux/mutex.h> | |
11 | #include <linux/spinlock.h> | |
12 | #include <linux/sysfs.h> | |
13 | #include <linux/timer.h> | |
14 | #include <linux/types.h> | |
15 | #include <linux/workqueue.h> | |
16 | ||
17 | #include <asm/atomic.h> | |
18 | #include <asm/byteorder.h> | |
19 | ||
20 | #define fw_notify(s, args...) printk(KERN_NOTICE KBUILD_MODNAME ": " s, ## args) | |
21 | #define fw_error(s, args...) printk(KERN_ERR KBUILD_MODNAME ": " s, ## args) | |
22 | ||
23 | static inline void fw_memcpy_from_be32(void *_dst, void *_src, size_t size) | |
24 | { | |
25 | u32 *dst = _dst; | |
26 | __be32 *src = _src; | |
27 | int i; | |
28 | ||
29 | for (i = 0; i < size / 4; i++) | |
30 | dst[i] = be32_to_cpu(src[i]); | |
31 | } | |
32 | ||
33 | static inline void fw_memcpy_to_be32(void *_dst, void *_src, size_t size) | |
34 | { | |
35 | fw_memcpy_from_be32(_dst, _src, size); | |
36 | } | |
37 | #define CSR_REGISTER_BASE 0xfffff0000000ULL | |
38 | ||
39 | /* register offsets are relative to CSR_REGISTER_BASE */ | |
40 | #define CSR_STATE_CLEAR 0x0 | |
41 | #define CSR_STATE_SET 0x4 | |
42 | #define CSR_NODE_IDS 0x8 | |
43 | #define CSR_RESET_START 0xc | |
44 | #define CSR_SPLIT_TIMEOUT_HI 0x18 | |
45 | #define CSR_SPLIT_TIMEOUT_LO 0x1c | |
46 | #define CSR_CYCLE_TIME 0x200 | |
47 | #define CSR_BUS_TIME 0x204 | |
48 | #define CSR_BUSY_TIMEOUT 0x210 | |
49 | #define CSR_BUS_MANAGER_ID 0x21c | |
50 | #define CSR_BANDWIDTH_AVAILABLE 0x220 | |
51 | #define CSR_CHANNELS_AVAILABLE 0x224 | |
52 | #define CSR_CHANNELS_AVAILABLE_HI 0x224 | |
53 | #define CSR_CHANNELS_AVAILABLE_LO 0x228 | |
54 | #define CSR_BROADCAST_CHANNEL 0x234 | |
55 | #define CSR_CONFIG_ROM 0x400 | |
56 | #define CSR_CONFIG_ROM_END 0x800 | |
57 | #define CSR_FCP_COMMAND 0xB00 | |
58 | #define CSR_FCP_RESPONSE 0xD00 | |
59 | #define CSR_FCP_END 0xF00 | |
60 | #define CSR_TOPOLOGY_MAP 0x1000 | |
61 | #define CSR_TOPOLOGY_MAP_END 0x1400 | |
62 | #define CSR_SPEED_MAP 0x2000 | |
63 | #define CSR_SPEED_MAP_END 0x3000 | |
64 | ||
65 | #define CSR_OFFSET 0x40 | |
66 | #define CSR_LEAF 0x80 | |
67 | #define CSR_DIRECTORY 0xc0 | |
68 | ||
69 | #define CSR_DESCRIPTOR 0x01 | |
70 | #define CSR_VENDOR 0x03 | |
71 | #define CSR_HARDWARE_VERSION 0x04 | |
72 | #define CSR_NODE_CAPABILITIES 0x0c | |
73 | #define CSR_UNIT 0x11 | |
74 | #define CSR_SPECIFIER_ID 0x12 | |
75 | #define CSR_VERSION 0x13 | |
76 | #define CSR_DEPENDENT_INFO 0x14 | |
77 | #define CSR_MODEL 0x17 | |
78 | #define CSR_INSTANCE 0x18 | |
79 | #define CSR_DIRECTORY_ID 0x20 | |
80 | ||
81 | struct fw_csr_iterator { | |
82 | u32 *p; | |
83 | u32 *end; | |
84 | }; | |
85 | ||
86 | void fw_csr_iterator_init(struct fw_csr_iterator *ci, u32 *p); | |
87 | int fw_csr_iterator_next(struct fw_csr_iterator *ci, int *key, int *value); | |
88 | ||
89 | extern struct bus_type fw_bus_type; | |
90 | ||
91 | struct fw_card_driver; | |
92 | struct fw_node; | |
93 | ||
94 | struct fw_card { | |
95 | const struct fw_card_driver *driver; | |
96 | struct device *device; | |
97 | struct kref kref; | |
98 | struct completion done; | |
99 | ||
100 | int node_id; | |
101 | int generation; | |
1e626fdc SR |
102 | int current_tlabel; |
103 | u64 tlabel_mask; | |
77c9a5da SR |
104 | struct list_head transaction_list; |
105 | struct timer_list flush_timer; | |
106 | unsigned long reset_jiffies; | |
107 | ||
108 | unsigned long long guid; | |
109 | unsigned max_receive; | |
110 | int link_speed; | |
111 | int config_rom_generation; | |
112 | ||
113 | spinlock_t lock; /* Take this lock when handling the lists in | |
114 | * this struct. */ | |
115 | struct fw_node *local_node; | |
116 | struct fw_node *root_node; | |
117 | struct fw_node *irm_node; | |
118 | u8 color; /* must be u8 to match the definition in struct fw_node */ | |
119 | int gap_count; | |
120 | bool beta_repeaters_present; | |
121 | ||
122 | int index; | |
123 | ||
124 | struct list_head link; | |
125 | ||
126 | /* Work struct for BM duties. */ | |
127 | struct delayed_work work; | |
128 | int bm_retries; | |
129 | int bm_generation; | |
6fdc0370 | 130 | __be32 bm_transaction_data[2]; |
77c9a5da SR |
131 | |
132 | bool broadcast_channel_allocated; | |
133 | u32 broadcast_channel; | |
134 | u32 topology_map[(CSR_TOPOLOGY_MAP_END - CSR_TOPOLOGY_MAP) / 4]; | |
135 | }; | |
136 | ||
77c9a5da SR |
137 | struct fw_attribute_group { |
138 | struct attribute_group *groups[2]; | |
139 | struct attribute_group group; | |
140 | struct attribute *attrs[12]; | |
141 | }; | |
142 | ||
143 | enum fw_device_state { | |
144 | FW_DEVICE_INITIALIZING, | |
145 | FW_DEVICE_RUNNING, | |
146 | FW_DEVICE_GONE, | |
147 | FW_DEVICE_SHUTDOWN, | |
148 | }; | |
149 | ||
150 | /* | |
151 | * Note, fw_device.generation always has to be read before fw_device.node_id. | |
152 | * Use SMP memory barriers to ensure this. Otherwise requests will be sent | |
153 | * to an outdated node_id if the generation was updated in the meantime due | |
154 | * to a bus reset. | |
155 | * | |
156 | * Likewise, fw-core will take care to update .node_id before .generation so | |
157 | * that whenever fw_device.generation is current WRT the actual bus generation, | |
158 | * fw_device.node_id is guaranteed to be current too. | |
159 | * | |
160 | * The same applies to fw_device.card->node_id vs. fw_device.generation. | |
161 | * | |
162 | * fw_device.config_rom and fw_device.config_rom_length may be accessed during | |
163 | * the lifetime of any fw_unit belonging to the fw_device, before device_del() | |
164 | * was called on the last fw_unit. Alternatively, they may be accessed while | |
165 | * holding fw_device_rwsem. | |
166 | */ | |
167 | struct fw_device { | |
168 | atomic_t state; | |
169 | struct fw_node *node; | |
170 | int node_id; | |
171 | int generation; | |
172 | unsigned max_speed; | |
173 | struct fw_card *card; | |
174 | struct device device; | |
175 | ||
176 | struct mutex client_list_mutex; | |
177 | struct list_head client_list; | |
178 | ||
179 | u32 *config_rom; | |
180 | size_t config_rom_length; | |
181 | int config_rom_retries; | |
182 | unsigned is_local:1; | |
837ec787 | 183 | unsigned max_rec:4; |
77c9a5da | 184 | unsigned cmc:1; |
837ec787 | 185 | unsigned irmc:1; |
77c9a5da SR |
186 | unsigned bc_implemented:2; |
187 | ||
188 | struct delayed_work work; | |
189 | struct fw_attribute_group attribute_group; | |
190 | }; | |
191 | ||
192 | static inline struct fw_device *fw_device(struct device *dev) | |
193 | { | |
194 | return container_of(dev, struct fw_device, device); | |
195 | } | |
196 | ||
197 | static inline int fw_device_is_shutdown(struct fw_device *device) | |
198 | { | |
199 | return atomic_read(&device->state) == FW_DEVICE_SHUTDOWN; | |
200 | } | |
201 | ||
202 | static inline struct fw_device *fw_device_get(struct fw_device *device) | |
203 | { | |
204 | get_device(&device->device); | |
205 | ||
206 | return device; | |
207 | } | |
208 | ||
209 | static inline void fw_device_put(struct fw_device *device) | |
210 | { | |
211 | put_device(&device->device); | |
212 | } | |
213 | ||
214 | int fw_device_enable_phys_dma(struct fw_device *device); | |
215 | ||
216 | /* | |
217 | * fw_unit.directory must not be accessed after device_del(&fw_unit.device). | |
218 | */ | |
219 | struct fw_unit { | |
220 | struct device device; | |
221 | u32 *directory; | |
222 | struct fw_attribute_group attribute_group; | |
223 | }; | |
224 | ||
225 | static inline struct fw_unit *fw_unit(struct device *dev) | |
226 | { | |
227 | return container_of(dev, struct fw_unit, device); | |
228 | } | |
229 | ||
230 | static inline struct fw_unit *fw_unit_get(struct fw_unit *unit) | |
231 | { | |
232 | get_device(&unit->device); | |
233 | ||
234 | return unit; | |
235 | } | |
236 | ||
237 | static inline void fw_unit_put(struct fw_unit *unit) | |
238 | { | |
239 | put_device(&unit->device); | |
240 | } | |
241 | ||
e5110d01 SR |
242 | static inline struct fw_device *fw_parent_device(struct fw_unit *unit) |
243 | { | |
244 | return fw_device(unit->device.parent); | |
245 | } | |
246 | ||
77c9a5da SR |
247 | struct ieee1394_device_id; |
248 | ||
249 | struct fw_driver { | |
250 | struct device_driver driver; | |
251 | /* Called when the parent device sits through a bus reset. */ | |
252 | void (*update)(struct fw_unit *unit); | |
253 | const struct ieee1394_device_id *id_table; | |
254 | }; | |
255 | ||
256 | struct fw_packet; | |
257 | struct fw_request; | |
258 | ||
259 | typedef void (*fw_packet_callback_t)(struct fw_packet *packet, | |
260 | struct fw_card *card, int status); | |
261 | typedef void (*fw_transaction_callback_t)(struct fw_card *card, int rcode, | |
262 | void *data, size_t length, | |
263 | void *callback_data); | |
264 | /* | |
265 | * Important note: The callback must guarantee that either fw_send_response() | |
266 | * or kfree() is called on the @request. | |
267 | */ | |
268 | typedef void (*fw_address_callback_t)(struct fw_card *card, | |
269 | struct fw_request *request, | |
270 | int tcode, int destination, int source, | |
271 | int generation, int speed, | |
272 | unsigned long long offset, | |
273 | void *data, size_t length, | |
274 | void *callback_data); | |
275 | ||
276 | struct fw_packet { | |
277 | int speed; | |
278 | int generation; | |
279 | u32 header[4]; | |
280 | size_t header_length; | |
281 | void *payload; | |
282 | size_t payload_length; | |
283 | dma_addr_t payload_bus; | |
284 | u32 timestamp; | |
285 | ||
286 | /* | |
287 | * This callback is called when the packet transmission has | |
288 | * completed; for successful transmission, the status code is | |
289 | * the ack received from the destination, otherwise it's a | |
290 | * negative errno: ENOMEM, ESTALE, ETIMEDOUT, ENODEV, EIO. | |
291 | * The callback can be called from tasklet context and thus | |
292 | * must never block. | |
293 | */ | |
294 | fw_packet_callback_t callback; | |
295 | int ack; | |
296 | struct list_head link; | |
297 | void *driver_data; | |
298 | }; | |
299 | ||
300 | struct fw_transaction { | |
301 | int node_id; /* The generation is implied; it is always the current. */ | |
302 | int tlabel; | |
303 | int timestamp; | |
304 | struct list_head link; | |
305 | ||
306 | struct fw_packet packet; | |
307 | ||
308 | /* | |
309 | * The data passed to the callback is valid only during the | |
310 | * callback. | |
311 | */ | |
312 | fw_transaction_callback_t callback; | |
313 | void *callback_data; | |
314 | }; | |
315 | ||
316 | struct fw_address_handler { | |
317 | u64 offset; | |
318 | size_t length; | |
319 | fw_address_callback_t address_callback; | |
320 | void *callback_data; | |
321 | struct list_head link; | |
322 | }; | |
323 | ||
324 | struct fw_address_region { | |
325 | u64 start; | |
326 | u64 end; | |
327 | }; | |
328 | ||
329 | extern const struct fw_address_region fw_high_memory_region; | |
330 | ||
331 | int fw_core_add_address_handler(struct fw_address_handler *handler, | |
332 | const struct fw_address_region *region); | |
333 | void fw_core_remove_address_handler(struct fw_address_handler *handler); | |
334 | void fw_send_response(struct fw_card *card, | |
335 | struct fw_request *request, int rcode); | |
336 | void fw_send_request(struct fw_card *card, struct fw_transaction *t, | |
337 | int tcode, int destination_id, int generation, int speed, | |
338 | unsigned long long offset, void *payload, size_t length, | |
339 | fw_transaction_callback_t callback, void *callback_data); | |
340 | int fw_cancel_transaction(struct fw_card *card, | |
341 | struct fw_transaction *transaction); | |
342 | int fw_run_transaction(struct fw_card *card, int tcode, int destination_id, | |
343 | int generation, int speed, unsigned long long offset, | |
344 | void *payload, size_t length); | |
345 | ||
c76acec6 JF |
346 | static inline int fw_stream_packet_destination_id(int tag, int channel, int sy) |
347 | { | |
348 | return tag << 14 | channel << 8 | sy; | |
349 | } | |
350 | ||
351 | struct fw_descriptor { | |
352 | struct list_head link; | |
353 | size_t length; | |
354 | u32 immediate; | |
355 | u32 key; | |
356 | const u32 *data; | |
357 | }; | |
358 | ||
359 | int fw_core_add_descriptor(struct fw_descriptor *desc); | |
360 | void fw_core_remove_descriptor(struct fw_descriptor *desc); | |
361 | ||
362 | /* | |
363 | * The iso packet format allows for an immediate header/payload part | |
364 | * stored in 'header' immediately after the packet info plus an | |
365 | * indirect payload part that is pointer to by the 'payload' field. | |
366 | * Applications can use one or the other or both to implement simple | |
367 | * low-bandwidth streaming (e.g. audio) or more advanced | |
368 | * scatter-gather streaming (e.g. assembling video frame automatically). | |
369 | */ | |
370 | struct fw_iso_packet { | |
371 | u16 payload_length; /* Length of indirect payload. */ | |
372 | u32 interrupt:1; /* Generate interrupt on this packet */ | |
373 | u32 skip:1; /* Set to not send packet at all. */ | |
374 | u32 tag:2; | |
375 | u32 sy:4; | |
376 | u32 header_length:8; /* Length of immediate header. */ | |
377 | u32 header[0]; | |
378 | }; | |
379 | ||
380 | #define FW_ISO_CONTEXT_TRANSMIT 0 | |
381 | #define FW_ISO_CONTEXT_RECEIVE 1 | |
382 | ||
383 | #define FW_ISO_CONTEXT_MATCH_TAG0 1 | |
384 | #define FW_ISO_CONTEXT_MATCH_TAG1 2 | |
385 | #define FW_ISO_CONTEXT_MATCH_TAG2 4 | |
386 | #define FW_ISO_CONTEXT_MATCH_TAG3 8 | |
387 | #define FW_ISO_CONTEXT_MATCH_ALL_TAGS 15 | |
388 | ||
389 | /* | |
390 | * An iso buffer is just a set of pages mapped for DMA in the | |
391 | * specified direction. Since the pages are to be used for DMA, they | |
392 | * are not mapped into the kernel virtual address space. We store the | |
393 | * DMA address in the page private. The helper function | |
394 | * fw_iso_buffer_map() will map the pages into a given vma. | |
395 | */ | |
396 | struct fw_iso_buffer { | |
397 | enum dma_data_direction direction; | |
398 | struct page **pages; | |
399 | int page_count; | |
400 | }; | |
401 | ||
402 | int fw_iso_buffer_init(struct fw_iso_buffer *buffer, struct fw_card *card, | |
403 | int page_count, enum dma_data_direction direction); | |
404 | void fw_iso_buffer_destroy(struct fw_iso_buffer *buffer, struct fw_card *card); | |
405 | ||
406 | struct fw_iso_context; | |
407 | typedef void (*fw_iso_callback_t)(struct fw_iso_context *context, | |
408 | u32 cycle, size_t header_length, | |
409 | void *header, void *data); | |
410 | struct fw_iso_context { | |
411 | struct fw_card *card; | |
412 | int type; | |
413 | int channel; | |
414 | int speed; | |
415 | size_t header_size; | |
416 | fw_iso_callback_t callback; | |
417 | void *callback_data; | |
418 | }; | |
419 | ||
420 | struct fw_iso_context *fw_iso_context_create(struct fw_card *card, | |
421 | int type, int channel, int speed, size_t header_size, | |
422 | fw_iso_callback_t callback, void *callback_data); | |
423 | int fw_iso_context_queue(struct fw_iso_context *ctx, | |
424 | struct fw_iso_packet *packet, | |
425 | struct fw_iso_buffer *buffer, | |
426 | unsigned long payload); | |
427 | int fw_iso_context_start(struct fw_iso_context *ctx, | |
428 | int cycle, int sync, int tags); | |
429 | int fw_iso_context_stop(struct fw_iso_context *ctx); | |
430 | void fw_iso_context_destroy(struct fw_iso_context *ctx); | |
431 | ||
77c9a5da | 432 | #endif /* _LINUX_FIREWIRE_H */ |