mirror of
git://source.winehq.org/git/wine.git
synced 2024-11-01 22:50:08 +00:00
cf08bd0e38
Signed-off-by: Jacek Caban <jacek@codeweavers.com> Signed-off-by: Alexandre Julliard <julliard@winehq.org>
802 lines
26 KiB
C
802 lines
26 KiB
C
/*
|
|
* Server-side device support
|
|
*
|
|
* Copyright (C) 2007 Alexandre Julliard
|
|
*
|
|
* This library is free software; you can redistribute it and/or
|
|
* modify it under the terms of the GNU Lesser General Public
|
|
* License as published by the Free Software Foundation; either
|
|
* version 2.1 of the License, or (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
* Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public
|
|
* License along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301, USA
|
|
*/
|
|
|
|
#include "config.h"
|
|
#include "wine/port.h"
|
|
|
|
#include <assert.h>
|
|
#include <fcntl.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <stdarg.h>
|
|
|
|
#include "ntstatus.h"
|
|
#define WIN32_NO_STATUS
|
|
#include "windef.h"
|
|
#include "winternl.h"
|
|
#include "ddk/wdm.h"
|
|
|
|
#include "object.h"
|
|
#include "file.h"
|
|
#include "handle.h"
|
|
#include "request.h"
|
|
#include "process.h"
|
|
|
|
/* IRP object */
|
|
|
|
struct irp_call
|
|
{
|
|
struct object obj; /* object header */
|
|
struct list dev_entry; /* entry in device queue */
|
|
struct list mgr_entry; /* entry in manager queue */
|
|
struct device_file *file; /* file containing this irp */
|
|
struct thread *thread; /* thread that queued the irp */
|
|
client_ptr_t user_arg; /* user arg used to identify the request */
|
|
struct async *async; /* pending async op */
|
|
irp_params_t params; /* irp parameters */
|
|
struct iosb *iosb; /* I/O status block */
|
|
};
|
|
|
|
static void irp_call_dump( struct object *obj, int verbose );
|
|
static int irp_call_signaled( struct object *obj, struct wait_queue_entry *entry );
|
|
static void irp_call_destroy( struct object *obj );
|
|
|
|
static const struct object_ops irp_call_ops =
|
|
{
|
|
sizeof(struct irp_call), /* size */
|
|
irp_call_dump, /* dump */
|
|
no_get_type, /* get_type */
|
|
add_queue, /* add_queue */
|
|
remove_queue, /* remove_queue */
|
|
irp_call_signaled, /* signaled */
|
|
no_satisfied, /* satisfied */
|
|
no_signal, /* signal */
|
|
no_get_fd, /* get_fd */
|
|
no_map_access, /* map_access */
|
|
default_get_sd, /* get_sd */
|
|
default_set_sd, /* set_sd */
|
|
no_lookup_name, /* lookup_name */
|
|
no_link_name, /* link_name */
|
|
NULL, /* unlink_name */
|
|
no_open_file, /* open_file */
|
|
no_close_handle, /* close_handle */
|
|
irp_call_destroy /* destroy */
|
|
};
|
|
|
|
|
|
/* device manager (a list of devices managed by the same client process) */
|
|
|
|
struct device_manager
|
|
{
|
|
struct object obj; /* object header */
|
|
struct list devices; /* list of devices */
|
|
struct list requests; /* list of pending irps across all devices */
|
|
};
|
|
|
|
static void device_manager_dump( struct object *obj, int verbose );
|
|
static int device_manager_signaled( struct object *obj, struct wait_queue_entry *entry );
|
|
static void device_manager_destroy( struct object *obj );
|
|
|
|
static const struct object_ops device_manager_ops =
|
|
{
|
|
sizeof(struct device_manager), /* size */
|
|
device_manager_dump, /* dump */
|
|
no_get_type, /* get_type */
|
|
add_queue, /* add_queue */
|
|
remove_queue, /* remove_queue */
|
|
device_manager_signaled, /* signaled */
|
|
no_satisfied, /* satisfied */
|
|
no_signal, /* signal */
|
|
no_get_fd, /* get_fd */
|
|
no_map_access, /* map_access */
|
|
default_get_sd, /* get_sd */
|
|
default_set_sd, /* set_sd */
|
|
no_lookup_name, /* lookup_name */
|
|
no_link_name, /* link_name */
|
|
NULL, /* unlink_name */
|
|
no_open_file, /* open_file */
|
|
no_close_handle, /* close_handle */
|
|
device_manager_destroy /* destroy */
|
|
};
|
|
|
|
|
|
/* device (a single device object) */
|
|
|
|
struct device
|
|
{
|
|
struct object obj; /* object header */
|
|
struct device_manager *manager; /* manager for this device (or NULL if deleted) */
|
|
char *unix_path; /* path to unix device if any */
|
|
client_ptr_t user_ptr; /* opaque ptr for client side */
|
|
struct list entry; /* entry in device manager list */
|
|
struct list files; /* list of open files */
|
|
};
|
|
|
|
static void device_dump( struct object *obj, int verbose );
|
|
static struct object_type *device_get_type( struct object *obj );
|
|
static void device_destroy( struct object *obj );
|
|
static struct object *device_open_file( struct object *obj, unsigned int access,
|
|
unsigned int sharing, unsigned int options );
|
|
|
|
static const struct object_ops device_ops =
|
|
{
|
|
sizeof(struct device), /* size */
|
|
device_dump, /* dump */
|
|
device_get_type, /* get_type */
|
|
no_add_queue, /* add_queue */
|
|
NULL, /* remove_queue */
|
|
NULL, /* signaled */
|
|
no_satisfied, /* satisfied */
|
|
no_signal, /* signal */
|
|
no_get_fd, /* get_fd */
|
|
default_fd_map_access, /* map_access */
|
|
default_get_sd, /* get_sd */
|
|
default_set_sd, /* set_sd */
|
|
no_lookup_name, /* lookup_name */
|
|
directory_link_name, /* link_name */
|
|
default_unlink_name, /* unlink_name */
|
|
device_open_file, /* open_file */
|
|
no_close_handle, /* close_handle */
|
|
device_destroy /* destroy */
|
|
};
|
|
|
|
|
|
/* device file (an open file handle to a device) */
|
|
|
|
struct device_file
|
|
{
|
|
struct object obj; /* object header */
|
|
struct device *device; /* device for this file */
|
|
struct fd *fd; /* file descriptor for irp */
|
|
client_ptr_t user_ptr; /* opaque ptr for client side */
|
|
struct list entry; /* entry in device list */
|
|
struct list requests; /* list of pending irp requests */
|
|
};
|
|
|
|
static void device_file_dump( struct object *obj, int verbose );
|
|
static struct fd *device_file_get_fd( struct object *obj );
|
|
static int device_file_close_handle( struct object *obj, struct process *process, obj_handle_t handle );
|
|
static void device_file_destroy( struct object *obj );
|
|
static enum server_fd_type device_file_get_fd_type( struct fd *fd );
|
|
static obj_handle_t device_file_read( struct fd *fd, const async_data_t *async_data, int blocking,
|
|
file_pos_t pos, struct iosb *iosb );
|
|
static obj_handle_t device_file_write( struct fd *fd, const async_data_t *async_data, int blocking,
|
|
file_pos_t pos, struct iosb *iosb );
|
|
static obj_handle_t device_file_flush( struct fd *fd, const async_data_t *async_data, int blocking );
|
|
static obj_handle_t device_file_ioctl( struct fd *fd, ioctl_code_t code, const async_data_t *async_data,
|
|
int blocking );
|
|
|
|
static const struct object_ops device_file_ops =
|
|
{
|
|
sizeof(struct device_file), /* size */
|
|
device_file_dump, /* dump */
|
|
no_get_type, /* get_type */
|
|
add_queue, /* add_queue */
|
|
remove_queue, /* remove_queue */
|
|
default_fd_signaled, /* signaled */
|
|
no_satisfied, /* satisfied */
|
|
no_signal, /* signal */
|
|
device_file_get_fd, /* get_fd */
|
|
default_fd_map_access, /* map_access */
|
|
default_get_sd, /* get_sd */
|
|
default_set_sd, /* set_sd */
|
|
no_lookup_name, /* lookup_name */
|
|
no_link_name, /* link_name */
|
|
NULL, /* unlink_name */
|
|
no_open_file, /* open_file */
|
|
device_file_close_handle, /* close_handle */
|
|
device_file_destroy /* destroy */
|
|
};
|
|
|
|
static const struct fd_ops device_file_fd_ops =
|
|
{
|
|
default_fd_get_poll_events, /* get_poll_events */
|
|
default_poll_event, /* poll_event */
|
|
device_file_get_fd_type, /* get_fd_type */
|
|
device_file_read, /* read */
|
|
device_file_write, /* write */
|
|
device_file_flush, /* flush */
|
|
device_file_ioctl, /* ioctl */
|
|
default_fd_queue_async, /* queue_async */
|
|
default_fd_reselect_async /* reselect_async */
|
|
};
|
|
|
|
|
|
static void irp_call_dump( struct object *obj, int verbose )
|
|
{
|
|
struct irp_call *irp = (struct irp_call *)obj;
|
|
fprintf( stderr, "IRP call file=%p\n", irp->file );
|
|
}
|
|
|
|
static int irp_call_signaled( struct object *obj, struct wait_queue_entry *entry )
|
|
{
|
|
struct irp_call *irp = (struct irp_call *)obj;
|
|
|
|
return !irp->file; /* file is cleared once the irp has completed */
|
|
}
|
|
|
|
static void irp_call_destroy( struct object *obj )
|
|
{
|
|
struct irp_call *irp = (struct irp_call *)obj;
|
|
|
|
if (irp->async)
|
|
{
|
|
async_terminate( irp->async, STATUS_CANCELLED );
|
|
release_object( irp->async );
|
|
}
|
|
if (irp->iosb) release_object( irp->iosb );
|
|
if (irp->file) release_object( irp->file );
|
|
if (irp->thread) release_object( irp->thread );
|
|
}
|
|
|
|
static struct irp_call *create_irp( struct device_file *file, const irp_params_t *params, struct iosb *iosb )
|
|
{
|
|
struct irp_call *irp;
|
|
|
|
if (!file->device->manager) /* it has been deleted */
|
|
{
|
|
set_error( STATUS_FILE_DELETED );
|
|
return NULL;
|
|
}
|
|
|
|
if ((irp = alloc_object( &irp_call_ops )))
|
|
{
|
|
irp->file = (struct device_file *)grab_object( file );
|
|
irp->thread = NULL;
|
|
irp->async = NULL;
|
|
irp->params = *params;
|
|
|
|
if (iosb)
|
|
{
|
|
irp->iosb = (struct iosb *)grab_object( iosb );
|
|
}
|
|
else if (!(irp->iosb = create_iosb( NULL, 0, 0 )))
|
|
{
|
|
release_object( irp );
|
|
irp = NULL;
|
|
}
|
|
}
|
|
return irp;
|
|
}
|
|
|
|
static void set_irp_result( struct irp_call *irp, unsigned int status,
|
|
const void *out_data, data_size_t out_size, data_size_t result )
|
|
{
|
|
struct device_file *file = irp->file;
|
|
struct iosb *iosb = irp->iosb;
|
|
|
|
if (!file) return; /* already finished */
|
|
|
|
/* FIXME: handle the STATUS_PENDING case */
|
|
iosb->status = status;
|
|
iosb->result = result;
|
|
iosb->out_size = min( iosb->out_size, out_size );
|
|
if (iosb->out_size && !(iosb->out_data = memdup( out_data, iosb->out_size )))
|
|
iosb->out_size = 0;
|
|
irp->file = NULL;
|
|
if (irp->async)
|
|
{
|
|
if (result) status = STATUS_ALERTED;
|
|
async_terminate( irp->async, status );
|
|
release_object( irp->async );
|
|
irp->async = NULL;
|
|
}
|
|
wake_up( &irp->obj, 0 );
|
|
|
|
/* remove it from the device queue */
|
|
list_remove( &irp->dev_entry );
|
|
release_object( irp ); /* no longer on the device queue */
|
|
release_object( file );
|
|
}
|
|
|
|
|
|
static void device_dump( struct object *obj, int verbose )
|
|
{
|
|
fputs( "Device\n", stderr );
|
|
}
|
|
|
|
static struct object_type *device_get_type( struct object *obj )
|
|
{
|
|
static const WCHAR name[] = {'D','e','v','i','c','e'};
|
|
static const struct unicode_str str = { name, sizeof(name) };
|
|
return get_object_type( &str );
|
|
}
|
|
|
|
static void device_destroy( struct object *obj )
|
|
{
|
|
struct device *device = (struct device *)obj;
|
|
|
|
assert( list_empty( &device->files ));
|
|
|
|
free( device->unix_path );
|
|
if (device->manager) list_remove( &device->entry );
|
|
}
|
|
|
|
static void add_irp_to_queue( struct device_file *file, struct irp_call *irp, struct thread *thread )
|
|
{
|
|
struct device_manager *manager = file->device->manager;
|
|
|
|
assert( manager );
|
|
|
|
grab_object( irp ); /* grab reference for queued irp */
|
|
irp->thread = thread ? (struct thread *)grab_object( thread ) : NULL;
|
|
list_add_tail( &file->requests, &irp->dev_entry );
|
|
list_add_tail( &manager->requests, &irp->mgr_entry );
|
|
if (list_head( &manager->requests ) == &irp->mgr_entry) wake_up( &manager->obj, 0 ); /* first one */
|
|
}
|
|
|
|
static struct object *device_open_file( struct object *obj, unsigned int access,
|
|
unsigned int sharing, unsigned int options )
|
|
{
|
|
struct device *device = (struct device *)obj;
|
|
struct device_file *file;
|
|
|
|
if (!(file = alloc_object( &device_file_ops ))) return NULL;
|
|
|
|
file->device = (struct device *)grab_object( device );
|
|
file->user_ptr = 0;
|
|
list_init( &file->requests );
|
|
list_add_tail( &device->files, &file->entry );
|
|
if (device->unix_path)
|
|
{
|
|
mode_t mode = 0666;
|
|
access = file->obj.ops->map_access( &file->obj, access );
|
|
file->fd = open_fd( NULL, device->unix_path, O_NONBLOCK | O_LARGEFILE,
|
|
&mode, access, sharing, options );
|
|
if (file->fd) set_fd_user( file->fd, &device_file_fd_ops, &file->obj );
|
|
}
|
|
else file->fd = alloc_pseudo_fd( &device_file_fd_ops, &file->obj, 0 );
|
|
|
|
if (!file->fd)
|
|
{
|
|
release_object( file );
|
|
return NULL;
|
|
}
|
|
|
|
allow_fd_caching( file->fd );
|
|
|
|
if (device->manager)
|
|
{
|
|
struct irp_call *irp;
|
|
irp_params_t params;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.create.major = IRP_MJ_CREATE;
|
|
params.create.access = access;
|
|
params.create.sharing = sharing;
|
|
params.create.options = options;
|
|
params.create.device = file->device->user_ptr;
|
|
|
|
if ((irp = create_irp( file, ¶ms, NULL )))
|
|
{
|
|
add_irp_to_queue( file, irp, NULL );
|
|
release_object( irp );
|
|
}
|
|
}
|
|
return &file->obj;
|
|
}
|
|
|
|
static void device_file_dump( struct object *obj, int verbose )
|
|
{
|
|
struct device_file *file = (struct device_file *)obj;
|
|
|
|
fprintf( stderr, "File on device %p\n", file->device );
|
|
}
|
|
|
|
static struct fd *device_file_get_fd( struct object *obj )
|
|
{
|
|
struct device_file *file = (struct device_file *)obj;
|
|
|
|
return (struct fd *)grab_object( file->fd );
|
|
}
|
|
|
|
static int device_file_close_handle( struct object *obj, struct process *process, obj_handle_t handle )
|
|
{
|
|
struct device_file *file = (struct device_file *)obj;
|
|
|
|
if (file->device->manager && obj->handle_count == 1) /* last handle */
|
|
{
|
|
struct irp_call *irp;
|
|
irp_params_t params;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.close.major = IRP_MJ_CLOSE;
|
|
params.close.file = file->user_ptr;
|
|
|
|
if ((irp = create_irp( file, ¶ms, NULL )))
|
|
{
|
|
add_irp_to_queue( file, irp, NULL );
|
|
release_object( irp );
|
|
}
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static void device_file_destroy( struct object *obj )
|
|
{
|
|
struct device_file *file = (struct device_file *)obj;
|
|
struct irp_call *irp, *next;
|
|
|
|
LIST_FOR_EACH_ENTRY_SAFE( irp, next, &file->requests, struct irp_call, dev_entry )
|
|
{
|
|
list_remove( &irp->dev_entry );
|
|
release_object( irp ); /* no longer on the device queue */
|
|
}
|
|
if (file->fd) release_object( file->fd );
|
|
list_remove( &file->entry );
|
|
release_object( file->device );
|
|
}
|
|
|
|
static void set_file_user_ptr( struct device_file *file, client_ptr_t ptr )
|
|
{
|
|
struct irp_call *irp;
|
|
|
|
if (file->user_ptr == ptr) return; /* nothing to do */
|
|
|
|
file->user_ptr = ptr;
|
|
|
|
/* update already queued irps */
|
|
|
|
LIST_FOR_EACH_ENTRY( irp, &file->requests, struct irp_call, dev_entry )
|
|
{
|
|
switch (irp->params.major)
|
|
{
|
|
case IRP_MJ_CLOSE: irp->params.close.file = ptr; break;
|
|
case IRP_MJ_READ: irp->params.read.file = ptr; break;
|
|
case IRP_MJ_WRITE: irp->params.write.file = ptr; break;
|
|
case IRP_MJ_FLUSH_BUFFERS: irp->params.flush.file = ptr; break;
|
|
case IRP_MJ_DEVICE_CONTROL: irp->params.ioctl.file = ptr; break;
|
|
}
|
|
}
|
|
}
|
|
|
|
/* queue an irp to the device */
|
|
static obj_handle_t queue_irp( struct device_file *file, struct irp_call *irp,
|
|
const async_data_t *async_data, int blocking )
|
|
{
|
|
obj_handle_t handle = 0;
|
|
|
|
if (blocking && !(handle = alloc_handle( current->process, irp, SYNCHRONIZE, 0 ))) return 0;
|
|
|
|
if (!(irp->async = fd_queue_async( file->fd, async_data, irp->iosb, ASYNC_TYPE_WAIT )))
|
|
{
|
|
if (handle) close_handle( current->process, handle );
|
|
return 0;
|
|
}
|
|
irp->user_arg = async_data->arg;
|
|
add_irp_to_queue( file, irp, current );
|
|
set_error( STATUS_PENDING );
|
|
return handle;
|
|
}
|
|
|
|
static enum server_fd_type device_file_get_fd_type( struct fd *fd )
|
|
{
|
|
return FD_TYPE_DEVICE;
|
|
}
|
|
|
|
static obj_handle_t device_file_read( struct fd *fd, const async_data_t *async_data, int blocking,
|
|
file_pos_t pos, struct iosb *iosb )
|
|
{
|
|
struct device_file *file = get_fd_user( fd );
|
|
struct irp_call *irp;
|
|
obj_handle_t handle;
|
|
irp_params_t params;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.read.major = IRP_MJ_READ;
|
|
params.read.key = 0;
|
|
params.read.pos = pos;
|
|
params.read.file = file->user_ptr;
|
|
|
|
irp = create_irp( file, ¶ms, iosb );
|
|
if (!irp) return 0;
|
|
|
|
handle = queue_irp( file, irp, async_data, blocking );
|
|
release_object( irp );
|
|
return handle;
|
|
}
|
|
|
|
static obj_handle_t device_file_write( struct fd *fd, const async_data_t *async_data, int blocking,
|
|
file_pos_t pos, struct iosb *iosb )
|
|
{
|
|
struct device_file *file = get_fd_user( fd );
|
|
struct irp_call *irp;
|
|
obj_handle_t handle;
|
|
irp_params_t params;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.write.major = IRP_MJ_WRITE;
|
|
params.write.key = 0;
|
|
params.write.pos = pos;
|
|
params.write.file = file->user_ptr;
|
|
|
|
irp = create_irp( file, ¶ms, iosb );
|
|
if (!irp) return 0;
|
|
|
|
handle = queue_irp( file, irp, async_data, blocking );
|
|
release_object( irp );
|
|
return handle;
|
|
}
|
|
|
|
static obj_handle_t device_file_flush( struct fd *fd, const async_data_t *async_data, int blocking )
|
|
{
|
|
struct device_file *file = get_fd_user( fd );
|
|
struct irp_call *irp;
|
|
obj_handle_t handle;
|
|
irp_params_t params;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.flush.major = IRP_MJ_FLUSH_BUFFERS;
|
|
params.flush.file = file->user_ptr;
|
|
|
|
irp = create_irp( file, ¶ms, NULL );
|
|
if (!irp) return 0;
|
|
|
|
handle = queue_irp( file, irp, async_data, blocking );
|
|
release_object( irp );
|
|
return handle;
|
|
}
|
|
|
|
static obj_handle_t device_file_ioctl( struct fd *fd, ioctl_code_t code, const async_data_t *async_data,
|
|
int blocking )
|
|
{
|
|
struct device_file *file = get_fd_user( fd );
|
|
struct irp_call *irp;
|
|
obj_handle_t handle;
|
|
irp_params_t params;
|
|
struct iosb *iosb;
|
|
|
|
memset( ¶ms, 0, sizeof(params) );
|
|
params.ioctl.major = IRP_MJ_DEVICE_CONTROL;
|
|
params.ioctl.code = code;
|
|
params.ioctl.file = file->user_ptr;
|
|
|
|
iosb = create_iosb( get_req_data(), get_req_data_size(), get_reply_max_size() );
|
|
if (!iosb) return 0;
|
|
irp = create_irp( file, ¶ms, iosb );
|
|
release_object(iosb);
|
|
if (!irp) return 0;
|
|
|
|
handle = queue_irp( file, irp, async_data, blocking );
|
|
release_object( irp );
|
|
return handle;
|
|
}
|
|
|
|
static struct device *create_device( struct object *root, const struct unicode_str *name,
|
|
struct device_manager *manager, unsigned int attr )
|
|
{
|
|
struct device *device;
|
|
|
|
if ((device = create_named_object( root, &device_ops, name, attr, NULL )))
|
|
{
|
|
if (get_error() != STATUS_OBJECT_NAME_EXISTS)
|
|
{
|
|
/* initialize it if it didn't already exist */
|
|
device->unix_path = NULL;
|
|
device->manager = manager;
|
|
list_add_tail( &manager->devices, &device->entry );
|
|
list_init( &device->files );
|
|
}
|
|
}
|
|
return device;
|
|
}
|
|
|
|
struct object *create_unix_device( struct object *root, const struct unicode_str *name,
|
|
const char *unix_path )
|
|
{
|
|
struct device *device;
|
|
|
|
if ((device = create_named_object( root, &device_ops, name, 0, NULL )))
|
|
{
|
|
device->unix_path = strdup( unix_path );
|
|
device->manager = NULL; /* no manager, requests go straight to the Unix device */
|
|
list_init( &device->files );
|
|
}
|
|
return &device->obj;
|
|
|
|
}
|
|
|
|
/* terminate requests when the underlying device is deleted */
|
|
static void delete_file( struct device_file *file )
|
|
{
|
|
struct irp_call *irp, *next;
|
|
|
|
/* terminate all pending requests */
|
|
LIST_FOR_EACH_ENTRY_SAFE( irp, next, &file->requests, struct irp_call, dev_entry )
|
|
{
|
|
list_remove( &irp->mgr_entry );
|
|
set_irp_result( irp, STATUS_FILE_DELETED, NULL, 0, 0 );
|
|
}
|
|
}
|
|
|
|
static void delete_device( struct device *device )
|
|
{
|
|
struct device_file *file, *next;
|
|
|
|
if (!device->manager) return; /* already deleted */
|
|
|
|
LIST_FOR_EACH_ENTRY_SAFE( file, next, &device->files, struct device_file, entry )
|
|
delete_file( file );
|
|
|
|
unlink_named_object( &device->obj );
|
|
list_remove( &device->entry );
|
|
device->manager = NULL;
|
|
}
|
|
|
|
|
|
static void device_manager_dump( struct object *obj, int verbose )
|
|
{
|
|
fprintf( stderr, "Device manager\n" );
|
|
}
|
|
|
|
static int device_manager_signaled( struct object *obj, struct wait_queue_entry *entry )
|
|
{
|
|
struct device_manager *manager = (struct device_manager *)obj;
|
|
|
|
return !list_empty( &manager->requests );
|
|
}
|
|
|
|
static void device_manager_destroy( struct object *obj )
|
|
{
|
|
struct device_manager *manager = (struct device_manager *)obj;
|
|
struct list *ptr;
|
|
|
|
while ((ptr = list_head( &manager->devices )))
|
|
{
|
|
struct device *device = LIST_ENTRY( ptr, struct device, entry );
|
|
delete_device( device );
|
|
}
|
|
}
|
|
|
|
static struct device_manager *create_device_manager(void)
|
|
{
|
|
struct device_manager *manager;
|
|
|
|
if ((manager = alloc_object( &device_manager_ops )))
|
|
{
|
|
list_init( &manager->devices );
|
|
list_init( &manager->requests );
|
|
}
|
|
return manager;
|
|
}
|
|
|
|
|
|
/* create a device manager */
|
|
DECL_HANDLER(create_device_manager)
|
|
{
|
|
struct device_manager *manager = create_device_manager();
|
|
|
|
if (manager)
|
|
{
|
|
reply->handle = alloc_handle( current->process, manager, req->access, req->attributes );
|
|
release_object( manager );
|
|
}
|
|
}
|
|
|
|
|
|
/* create a device */
|
|
DECL_HANDLER(create_device)
|
|
{
|
|
struct device *device;
|
|
struct unicode_str name = get_req_unicode_str();
|
|
struct device_manager *manager;
|
|
struct object *root = NULL;
|
|
|
|
if (!(manager = (struct device_manager *)get_handle_obj( current->process, req->manager,
|
|
0, &device_manager_ops )))
|
|
return;
|
|
|
|
if (req->rootdir && !(root = get_directory_obj( current->process, req->rootdir )))
|
|
{
|
|
release_object( manager );
|
|
return;
|
|
}
|
|
|
|
if ((device = create_device( root, &name, manager, req->attributes )))
|
|
{
|
|
device->user_ptr = req->user_ptr;
|
|
reply->handle = alloc_handle( current->process, device, req->access, req->attributes );
|
|
release_object( device );
|
|
}
|
|
|
|
if (root) release_object( root );
|
|
release_object( manager );
|
|
}
|
|
|
|
|
|
/* delete a device */
|
|
DECL_HANDLER(delete_device)
|
|
{
|
|
struct device *device;
|
|
|
|
if ((device = (struct device *)get_handle_obj( current->process, req->handle, 0, &device_ops )))
|
|
{
|
|
delete_device( device );
|
|
release_object( device );
|
|
}
|
|
}
|
|
|
|
|
|
/* retrieve the next pending device irp request */
|
|
DECL_HANDLER(get_next_device_request)
|
|
{
|
|
struct irp_call *irp;
|
|
struct device_manager *manager;
|
|
struct list *ptr;
|
|
struct iosb *iosb;
|
|
|
|
reply->params.major = IRP_MJ_MAXIMUM_FUNCTION + 1;
|
|
|
|
if (!(manager = (struct device_manager *)get_handle_obj( current->process, req->manager,
|
|
0, &device_manager_ops )))
|
|
return;
|
|
|
|
if (req->prev)
|
|
{
|
|
if ((irp = (struct irp_call *)get_handle_obj( current->process, req->prev, 0, &irp_call_ops )))
|
|
{
|
|
set_irp_result( irp, req->status, NULL, 0, 0 );
|
|
close_handle( current->process, req->prev ); /* avoid an extra round-trip for close */
|
|
release_object( irp );
|
|
}
|
|
clear_error();
|
|
}
|
|
|
|
if ((ptr = list_head( &manager->requests )))
|
|
{
|
|
irp = LIST_ENTRY( ptr, struct irp_call, mgr_entry );
|
|
if (irp->thread)
|
|
{
|
|
reply->client_pid = get_process_id( irp->thread->process );
|
|
reply->client_tid = get_thread_id( irp->thread );
|
|
}
|
|
reply->params = irp->params;
|
|
iosb = irp->iosb;
|
|
reply->in_size = iosb->in_size;
|
|
reply->out_size = iosb->out_size;
|
|
if (iosb->in_size > get_reply_max_size()) set_error( STATUS_BUFFER_OVERFLOW );
|
|
else if ((reply->next = alloc_handle( current->process, irp, 0, 0 )))
|
|
{
|
|
set_reply_data_ptr( iosb->in_data, iosb->in_size );
|
|
iosb->in_data = NULL;
|
|
iosb->in_size = 0;
|
|
list_remove( &irp->mgr_entry );
|
|
list_init( &irp->mgr_entry );
|
|
}
|
|
}
|
|
else set_error( STATUS_PENDING );
|
|
|
|
release_object( manager );
|
|
}
|
|
|
|
|
|
/* store results of an async irp */
|
|
DECL_HANDLER(set_irp_result)
|
|
{
|
|
struct irp_call *irp;
|
|
|
|
if ((irp = (struct irp_call *)get_handle_obj( current->process, req->handle, 0, &irp_call_ops )))
|
|
{
|
|
if (irp->file) set_file_user_ptr( irp->file, req->file_ptr );
|
|
set_irp_result( irp, req->status, get_req_data(), get_req_data_size(), req->size );
|
|
close_handle( current->process, req->handle ); /* avoid an extra round-trip for close */
|
|
release_object( irp );
|
|
}
|
|
}
|