blob: fb0321300cce4a01b646e77a5f0d040c0f46f25d [file] [log] [blame]
Stefan Richter77c9a5d2009-06-05 16:26:18 +02001#ifndef _FIREWIRE_CORE_H
2#define _FIREWIRE_CORE_H
3
Stefan Richter77c9a5d2009-06-05 16:26:18 +02004#include <linux/fs.h>
5#include <linux/list.h>
6#include <linux/idr.h>
7#include <linux/mm_types.h>
8#include <linux/rwsem.h>
9#include <linux/slab.h>
10#include <linux/types.h>
11
12#include <asm/atomic.h>
13
14struct device;
15struct fw_card;
16struct fw_device;
17struct fw_iso_buffer;
18struct fw_iso_context;
19struct fw_iso_packet;
20struct fw_node;
21struct fw_packet;
22
23
24/* -card */
25
26/* bitfields within the PHY registers */
27#define PHY_LINK_ACTIVE 0x80
28#define PHY_CONTENDER 0x40
29#define PHY_BUS_RESET 0x40
30#define PHY_BUS_SHORT_RESET 0x40
31
32#define BANDWIDTH_AVAILABLE_INITIAL 4915
33#define BROADCAST_CHANNEL_INITIAL (1 << 31 | 31)
34#define BROADCAST_CHANNEL_VALID (1 << 30)
35
36struct fw_card_driver {
37 /*
38 * Enable the given card with the given initial config rom.
39 * This function is expected to activate the card, and either
40 * enable the PHY or set the link_on bit and initiate a bus
41 * reset.
42 */
Stefan Richter8e859732009-10-08 00:41:59 +020043 int (*enable)(struct fw_card *card,
44 const __be32 *config_rom, size_t length);
Stefan Richter77c9a5d2009-06-05 16:26:18 +020045
46 int (*update_phy_reg)(struct fw_card *card, int address,
47 int clear_bits, int set_bits);
48
49 /*
50 * Update the config rom for an enabled card. This function
51 * should change the config rom that is presented on the bus
Stefan Richter8e859732009-10-08 00:41:59 +020052 * and initiate a bus reset.
Stefan Richter77c9a5d2009-06-05 16:26:18 +020053 */
54 int (*set_config_rom)(struct fw_card *card,
Stefan Richter8e859732009-10-08 00:41:59 +020055 const __be32 *config_rom, size_t length);
Stefan Richter77c9a5d2009-06-05 16:26:18 +020056
57 void (*send_request)(struct fw_card *card, struct fw_packet *packet);
58 void (*send_response)(struct fw_card *card, struct fw_packet *packet);
59 /* Calling cancel is valid once a packet has been submitted. */
60 int (*cancel_packet)(struct fw_card *card, struct fw_packet *packet);
61
62 /*
63 * Allow the specified node ID to do direct DMA out and in of
64 * host memory. The card will disable this for all node when
65 * a bus reset happens, so driver need to reenable this after
66 * bus reset. Returns 0 on success, -ENODEV if the card
67 * doesn't support this, -ESTALE if the generation doesn't
68 * match.
69 */
70 int (*enable_phys_dma)(struct fw_card *card,
71 int node_id, int generation);
72
Stefan Richter168cf9a2010-02-14 18:49:18 +010073 u32 (*get_cycle_time)(struct fw_card *card);
Stefan Richter77c9a5d2009-06-05 16:26:18 +020074
75 struct fw_iso_context *
76 (*allocate_iso_context)(struct fw_card *card,
77 int type, int channel, size_t header_size);
78 void (*free_iso_context)(struct fw_iso_context *ctx);
79
80 int (*start_iso)(struct fw_iso_context *ctx,
81 s32 cycle, u32 sync, u32 tags);
82
83 int (*queue_iso)(struct fw_iso_context *ctx,
84 struct fw_iso_packet *packet,
85 struct fw_iso_buffer *buffer,
86 unsigned long payload);
87
88 int (*stop_iso)(struct fw_iso_context *ctx);
89};
90
91void fw_card_initialize(struct fw_card *card,
92 const struct fw_card_driver *driver, struct device *device);
93int fw_card_add(struct fw_card *card,
94 u32 max_receive, u32 link_speed, u64 guid);
95void fw_core_remove_card(struct fw_card *card);
96int fw_core_initiate_bus_reset(struct fw_card *card, int short_reset);
Stefan Richtercb7c96d2009-10-08 00:42:53 +020097int fw_compute_block_crc(__be32 *block);
Stefan Richter77c9a5d2009-06-05 16:26:18 +020098void fw_schedule_bm_work(struct fw_card *card, unsigned long delay);
99
Stefan Richter18668ff2009-09-06 18:49:48 +0200100static inline struct fw_card *fw_card_get(struct fw_card *card)
101{
102 kref_get(&card->kref);
103
104 return card;
105}
106
107void fw_card_release(struct kref *kref);
108
109static inline void fw_card_put(struct fw_card *card)
110{
111 kref_put(&card->kref, fw_card_release);
112}
113
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200114
115/* -cdev */
116
117extern const struct file_operations fw_device_ops;
118
119void fw_device_cdev_update(struct fw_device *device);
120void fw_device_cdev_remove(struct fw_device *device);
121
122
123/* -device */
124
125extern struct rw_semaphore fw_device_rwsem;
126extern struct idr fw_device_idr;
127extern int fw_cdev_major;
128
129struct fw_device *fw_device_get_by_devt(dev_t devt);
Stefan Richter099d5412009-06-06 18:37:25 +0200130int fw_device_set_broadcast_channel(struct device *dev, void *gen);
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200131void fw_node_event(struct fw_card *card, struct fw_node *node, int event);
132
133
134/* -iso */
135
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200136int fw_iso_buffer_map(struct fw_iso_buffer *buffer, struct vm_area_struct *vma);
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200137void fw_iso_resource_manage(struct fw_card *card, int generation,
Stefan Richter6fdc0372009-06-20 13:23:59 +0200138 u64 channels_mask, int *channel, int *bandwidth,
139 bool allocate, __be32 buffer[2]);
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200140
141
142/* -topology */
143
144enum {
145 FW_NODE_CREATED,
146 FW_NODE_UPDATED,
147 FW_NODE_DESTROYED,
148 FW_NODE_LINK_ON,
149 FW_NODE_LINK_OFF,
150 FW_NODE_INITIATED_RESET,
151};
152
153struct fw_node {
154 u16 node_id;
155 u8 color;
156 u8 port_count;
157 u8 link_on:1;
158 u8 initiated_reset:1;
159 u8 b_path:1;
160 u8 phy_speed:2; /* As in the self ID packet. */
161 u8 max_speed:2; /* Minimum of all phy-speeds on the path from the
162 * local node to this node. */
163 u8 max_depth:4; /* Maximum depth to any leaf node */
164 u8 max_hops:4; /* Max hops in this sub tree */
165 atomic_t ref_count;
166
167 /* For serializing node topology into a list. */
168 struct list_head link;
169
170 /* Upper layer specific data. */
171 void *data;
172
173 struct fw_node *ports[0];
174};
175
176static inline struct fw_node *fw_node_get(struct fw_node *node)
177{
178 atomic_inc(&node->ref_count);
179
180 return node;
181}
182
183static inline void fw_node_put(struct fw_node *node)
184{
185 if (atomic_dec_and_test(&node->ref_count))
186 kfree(node);
187}
188
189void fw_core_handle_bus_reset(struct fw_card *card, int node_id,
190 int generation, int self_id_count, u32 *self_ids);
191void fw_destroy_nodes(struct fw_card *card);
192
193/*
194 * Check whether new_generation is the immediate successor of old_generation.
195 * Take counter roll-over at 255 (as per OHCI) into account.
196 */
197static inline bool is_next_generation(int new_generation, int old_generation)
198{
199 return (new_generation & 0xff) == ((old_generation + 1) & 0xff);
200}
201
202
203/* -transaction */
204
205#define TCODE_IS_READ_REQUEST(tcode) (((tcode) & ~1) == 4)
206#define TCODE_IS_BLOCK_PACKET(tcode) (((tcode) & 1) != 0)
207#define TCODE_IS_REQUEST(tcode) (((tcode) & 2) == 0)
208#define TCODE_IS_RESPONSE(tcode) (((tcode) & 2) != 0)
209#define TCODE_HAS_REQUEST_DATA(tcode) (((tcode) & 12) != 4)
210#define TCODE_HAS_RESPONSE_DATA(tcode) (((tcode) & 12) != 0)
211
212#define LOCAL_BUS 0xffc0
213
214void fw_core_handle_request(struct fw_card *card, struct fw_packet *request);
215void fw_core_handle_response(struct fw_card *card, struct fw_packet *packet);
216void fw_fill_response(struct fw_packet *response, u32 *request_header,
217 int rcode, void *payload, size_t length);
218void fw_flush_transactions(struct fw_card *card);
219void fw_send_phy_config(struct fw_card *card,
220 int node_id, int generation, int gap_count);
221
Stefan Richter77c9a5d2009-06-05 16:26:18 +0200222#endif /* _FIREWIRE_CORE_H */