struct ltt_channel_setting, kref);
struct ltt_channel_setting *iter;
- if (atomic_read(&index_kref.refcount) == 0
- && atomic_read(&setting->kref.refcount) == 0) {
+ if (uatomic_read(&index_kref.refcount) == 0
+ && uatomic_read(&setting->kref.refcount) == 0) {
list_del(&setting->list);
kfree(setting);
mutex_lock(<t_channel_mutex);
setting = lookup_channel(name);
if (setting) {
- if (atomic_read(&setting->kref.refcount) == 0)
+ if (uatomic_read(&setting->kref.refcount) == 0)
goto init_kref;
else {
kref_get(&setting->kref);
mutex_lock(<t_channel_mutex);
setting = lookup_channel(name);
- if (!setting || atomic_read(&setting->kref.refcount) == 0) {
+ if (!setting || uatomic_read(&setting->kref.refcount) == 0) {
ret = -ENOENT;
goto end;
}
mutex_lock(<t_channel_mutex);
setting = lookup_channel(name);
- if (!setting || atomic_read(&setting->kref.refcount) == 0) {
+ if (!setting || uatomic_read(&setting->kref.refcount) == 0) {
ret = -ENOENT;
goto end;
}
struct ltt_channel_setting *iter;
list_for_each_entry(iter, <t_channels, list)
- if (iter->index == index && atomic_read(&iter->kref.refcount))
+ if (iter->index == index && uatomic_read(&iter->kref.refcount))
return iter->name;
return NULL;
}
list_for_each_entry(iter, <t_channels, list)
if (!strcmp(iter->name, name)
- && atomic_read(&iter->kref.refcount))
+ && uatomic_read(&iter->kref.refcount))
return iter;
return NULL;
}
* Called with trace lock held. Does not perform the trace buffer allocation,
* because we must let the user overwrite specific channel sizes.
*/
-struct ltt_channel_struct *ltt_channels_trace_alloc(unsigned int *nr_channels,
+struct ust_channel *ltt_channels_trace_alloc(unsigned int *nr_channels,
int overwrite,
int active)
{
- struct ltt_channel_struct *channel = NULL;
+ struct ust_channel *channel = NULL;
struct ltt_channel_setting *iter;
mutex_lock(<t_channel_mutex);
WARN("ltt_channels_trace_alloc: no free_index; are there any probes connected?");
goto end;
}
- if (!atomic_read(&index_kref.refcount))
+ if (!uatomic_read(&index_kref.refcount))
kref_init(&index_kref);
else
kref_get(&index_kref);
*nr_channels = free_index;
- channel = kzalloc(sizeof(struct ltt_channel_struct) * free_index,
+ channel = kzalloc(sizeof(struct ust_channel) * free_index,
GFP_KERNEL);
if (!channel) {
WARN("ltt_channel_struct: channel null after alloc");
goto end;
}
list_for_each_entry(iter, <t_channels, list) {
- if (!atomic_read(&iter->kref.refcount))
+ if (!uatomic_read(&iter->kref.refcount))
continue;
channel[iter->index].subbuf_size = iter->subbuf_size;
channel[iter->index].subbuf_cnt = iter->subbuf_cnt;
* Called with trace lock held. The actual channel buffers must be freed before
* this function is called.
*/
-void ltt_channels_trace_free(struct ltt_channel_struct *channels)
+void ltt_channels_trace_free(struct ust_channel *channels)
{
lock_markers();
mutex_lock(<t_channel_mutex);