forked from KlausKnopper/cloop
-
Notifications
You must be signed in to change notification settings - Fork 0
/
cloop.c
1096 lines (1029 loc) · 34.7 KB
/
cloop.c
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
/*
* compressed_loop.c: Read-only compressed loop blockdevice
* hacked up by Rusty in 1999, extended and maintained by Klaus Knopper
*
* A cloop file looks like this:
* [32-bit uncompressed block size: network order]
* [32-bit number of blocks (n_blocks): network order]
* [64-bit file offsets of start of blocks: network order]
* ...
* (n_blocks + 1).
* n_blocks consisting of:
* [compressed block]
*
* Every version greatly inspired by code seen in loop.c
* by Theodore Ts'o, 3/29/93.
*
* Copyright 1999-2009 by Paul `Rusty' Russell & Klaus Knopper.
* Redistribution of this file is permitted under the GNU Public License.
*
*/
#define CLOOP_NAME "cloop"
#define CLOOP_VERSION "3.14"
#define CLOOP_MAX 8
#ifndef KBUILD_MODNAME
#define KBUILD_MODNAME cloop
#endif
#ifndef KBUILD_BASENAME
#define KBUILD_BASENAME cloop
#endif
#include <linux/kernel.h>
#include <linux/version.h>
#include <linux/module.h>
#include <linux/init.h>
#include <linux/sched.h>
#include <linux/fs.h>
#include <linux/file.h>
#include <linux/stat.h>
#include <linux/errno.h>
#include <linux/major.h>
#include <linux/vmalloc.h>
#include <linux/slab.h>
#include <linux/mutex.h>
#include <asm/div64.h> /* do_div() for 64bit division */
#include <asm/uaccess.h>
#include <asm/byteorder.h>
/* Use zlib_inflate from lib/zlib_inflate */
#include <linux/zutil.h>
#include <linux/loop.h>
#include <linux/kthread.h>
#include <linux/compat.h>
#include "cloop.h"
/* New License scheme */
#ifdef MODULE_LICENSE
MODULE_LICENSE("GPL");
#endif
#ifdef MODULE_AUTHOR
MODULE_AUTHOR("Klaus Knopper (current maintainer), Paul Russel (initial Kernel 2.2 version)");
#endif
#ifdef MODULE_DESCRIPTION
MODULE_DESCRIPTION("Transparently decompressing loopback block device");
#endif
#ifndef MIN
#define MIN(x,y) ((x) < (y) ? (x) : (y))
#endif
#ifndef MAX
#define MAX(x,y) ((x) > (y) ? (x) : (y))
#endif
/* Use experimental major for now */
#define MAJOR_NR 240
/* #define DEVICE_NAME CLOOP_NAME */
/* #define DEVICE_NR(device) (MINOR(device)) */
/* #define DEVICE_ON(device) */
/* #define DEVICE_OFF(device) */
/* #define DEVICE_NO_RANDOM */
/* #define TIMEOUT_VALUE (6 * HZ) */
#include <linux/blkdev.h>
#include <linux/buffer_head.h>
#if 0
#define DEBUGP printk
#else
#define DEBUGP(format, x...)
#endif
/* One file can be opened at module insertion time */
/* insmod cloop file=/path/to/file */
static char *file=NULL;
static unsigned int preload=0;
static unsigned int cloop_max=CLOOP_MAX;
module_param(file, charp, 0);
module_param(preload, uint, 0);
module_param(cloop_max, uint, 0);
MODULE_PARM_DESC(file, "Initial cloop image file (full path) for /dev/cloop");
MODULE_PARM_DESC(preload, "Preload n blocks of cloop data into memory");
MODULE_PARM_DESC(cloop_max, "Maximum number of cloop devices (default 8)");
static struct file *initial_file=NULL;
static int cloop_major=MAJOR_NR;
/* Number of buffered decompressed blocks */
#define BUFFERED_BLOCKS 8
struct cloop_device
{
/* Copied straight from the file */
struct cloop_head head;
/* An array of offsets of compressed blocks within the file */
loff_t *offsets;
/* We buffer some uncompressed blocks for performance */
int buffered_blocknum[BUFFERED_BLOCKS];
int current_bufnum;
void *buffer[BUFFERED_BLOCKS];
void *compressed_buffer;
size_t preload_array_size; /* Size of pointer array in blocks */
size_t preload_size; /* Number of successfully allocated blocks */
char **preload_cache; /* Pointers to preloaded blocks */
z_stream zstream;
struct file *backing_file; /* associated file */
struct inode *backing_inode; /* for bmap */
unsigned long largest_block;
unsigned int underlying_blksize;
int clo_number;
int refcnt;
struct block_device *bdev;
int isblkdev;
/* Lock for kernel block device queue */
spinlock_t queue_lock;
/* mutex for ioctl() */
struct mutex clo_ctl_mutex;
struct list_head clo_list;
struct task_struct *clo_thread;
wait_queue_head_t clo_event;
struct request_queue *clo_queue;
struct gendisk *clo_disk;
int suspended;
char clo_file_name[LO_NAME_SIZE];
};
/* Changed in 2.639: cloop_dev is now a an array of cloop_dev pointers,
so we can specify how many devices we need via parameters. */
static struct cloop_device **cloop_dev;
static const char *cloop_name=CLOOP_NAME;
static int cloop_count = 0;
#if (!(defined(CONFIG_ZLIB_INFLATE) || defined(CONFIG_ZLIB_INFLATE_MODULE))) /* Must be compiled into kernel. */
#error "Invalid Kernel configuration. CONFIG_ZLIB_INFLATE support is needed for cloop."
#endif
/* Use __get_free_pages instead of vmalloc, allows up to 32 pages,
* 2MB in one piece */
static void *cloop_malloc(size_t size)
{
/* kmalloc will fail after the system is running for a while, */
/* when large orders can't return contiguous memory. */
/* Let's just use vmalloc for now. :-/ */
/* int order = get_order(size); */
/* if(order <= KMALLOC_MAX_ORDER) */
/* return (void *)kmalloc(size, GFP_KERNEL); */
/* else if(order < MAX_ORDER) */
/* return (void *)__get_free_pages(GFP_KERNEL, order); */
return (void *)vmalloc(size);
}
static void cloop_free(void *mem, size_t size)
{
/* int order = get_order(size); */
/* if(order <= KMALLOC_MAX_ORDER) */
/* kfree(mem); */
/* else if(order < MAX_ORDER) */
/* free_pages((unsigned long)mem, order); */
/* else */
vfree(mem);
}
static int uncompress(struct cloop_device *clo,
unsigned char *dest, unsigned long *destLen,
unsigned char *source, unsigned long sourceLen)
{
/* Most of this code can be found in fs/cramfs/uncompress.c */
int err;
clo->zstream.next_in = source;
clo->zstream.avail_in = sourceLen;
clo->zstream.next_out = dest;
clo->zstream.avail_out = *destLen;
err = zlib_inflateReset(&clo->zstream);
if (err != Z_OK)
{
printk(KERN_ERR "%s: zlib_inflateReset error %d\n", cloop_name, err);
zlib_inflateEnd(&clo->zstream); zlib_inflateInit(&clo->zstream);
}
err = zlib_inflate(&clo->zstream, Z_FINISH);
*destLen = clo->zstream.total_out;
if (err != Z_STREAM_END) return err;
return Z_OK;
}
static ssize_t cloop_read_from_file(struct cloop_device *clo, struct file *f, char *buf,
loff_t pos, size_t buf_len)
{
size_t buf_done=0;
while (buf_done < buf_len)
{
size_t size = buf_len - buf_done, size_read;
/* kernel_read() only supports 32 bit offsets, so we use vfs_read() instead. */
/* int size_read = kernel_read(f, pos, buf + buf_done, size); */
mm_segment_t old_fs = get_fs();
set_fs(get_ds());
size_read = vfs_read(f, (void __user *)(buf + buf_done), size, &pos);
set_fs(old_fs);
if(size_read <= 0)
{
printk(KERN_ERR "%s: Read error %d at pos %Lu in file %s, "
"%d bytes lost.\n", cloop_name, (int)size_read, pos,
file, (int)size);
memset(buf + buf_len - size, 0, size);
break;
}
buf_done += size_read;
}
return buf_done;
}
/* This looks more complicated than it is */
/* Returns number of block buffer to use for this request */
static int cloop_load_buffer(struct cloop_device *clo, int blocknum)
{
unsigned int buf_done = 0;
unsigned long buflen;
unsigned int buf_length;
int ret;
int i;
if(blocknum > ntohl(clo->head.num_blocks) || blocknum < 0)
{
printk(KERN_WARNING "%s: Invalid block number %d requested.\n",
cloop_name, blocknum);
return -1;
}
/* Quick return if the block we seek is already in one of the buffers. */
/* Return number of buffer */
for(i=0; i<BUFFERED_BLOCKS; i++)
if (blocknum == clo->buffered_blocknum[i])
{
DEBUGP(KERN_INFO "cloop_load_buffer: Found buffered block %d\n", i);
return i;
}
buf_length = be64_to_cpu(clo->offsets[blocknum+1]) - be64_to_cpu(clo->offsets[blocknum]);
/* Load one compressed block from the file. */
cloop_read_from_file(clo, clo->backing_file, (char *)clo->compressed_buffer,
be64_to_cpu(clo->offsets[blocknum]), buf_length);
buflen = ntohl(clo->head.block_size);
/* Go to next position in the block ring buffer */
clo->current_bufnum++;
if(clo->current_bufnum >= BUFFERED_BLOCKS) clo->current_bufnum = 0;
/* Do the uncompression */
ret = uncompress(clo, clo->buffer[clo->current_bufnum], &buflen, clo->compressed_buffer,
buf_length);
/* DEBUGP("cloop: buflen after uncompress: %ld\n",buflen); */
if (ret != 0)
{
printk(KERN_ERR "%s: zlib decompression error %i uncompressing block %u %u/%lu/%u/%u "
"%Lu-%Lu\n", cloop_name, ret, blocknum,
ntohl(clo->head.block_size), buflen, buf_length, buf_done,
be64_to_cpu(clo->offsets[blocknum]), be64_to_cpu(clo->offsets[blocknum+1]));
clo->buffered_blocknum[clo->current_bufnum] = -1;
return -1;
}
clo->buffered_blocknum[clo->current_bufnum] = blocknum;
return clo->current_bufnum;
}
/* This function does all the real work. */
/* returns "uptodate" */
static int cloop_handle_request(struct cloop_device *clo, struct request *req)
{
int buffered_blocknum = -1;
int preloaded = 0;
loff_t offset = (loff_t) blk_rq_pos(req)<<9; /* req->sector<<9 */
struct bio_vec bvec;
struct req_iterator iter;
rq_for_each_segment(bvec, req, iter)
{
unsigned long len = bvec.bv_len;
char *to_ptr = kmap(bvec.bv_page) + bvec.bv_offset;
while(len > 0)
{
u_int32_t length_in_buffer;
loff_t block_offset = offset;
u_int32_t offset_in_buffer;
char *from_ptr;
/* do_div (div64.h) returns the 64bit division remainder and */
/* puts the result in the first argument, i.e. block_offset */
/* becomes the blocknumber to load, and offset_in_buffer the */
/* position in the buffer */
offset_in_buffer = do_div(block_offset, ntohl(clo->head.block_size));
/* Lookup preload cache */
if(block_offset < clo->preload_size && clo->preload_cache != NULL &&
clo->preload_cache[block_offset] != NULL)
{ /* Copy from cache */
preloaded = 1;
from_ptr = clo->preload_cache[block_offset];
}
else
{
preloaded = 0;
buffered_blocknum = cloop_load_buffer(clo,block_offset);
if(buffered_blocknum == -1) break; /* invalid data, leave inner loop */
/* Copy from buffer */
from_ptr = clo->buffer[buffered_blocknum];
}
/* Now, at least part of what we want will be in the buffer. */
length_in_buffer = ntohl(clo->head.block_size) - offset_in_buffer;
if(length_in_buffer > len)
{
/* DEBUGP("Warning: length_in_buffer=%u > len=%u\n",
length_in_buffer,len); */
length_in_buffer = len;
}
memcpy(to_ptr, from_ptr + offset_in_buffer, length_in_buffer);
to_ptr += length_in_buffer;
len -= length_in_buffer;
offset += length_in_buffer;
} /* while inner loop */
kunmap(bvec.bv_page);
} /* end rq_for_each_segment*/
return ((buffered_blocknum!=-1) || preloaded);
}
/* Adopted from loop.c, a kernel thread to handle physical reads and
* decompression. */
static int cloop_thread(void *data)
{
struct cloop_device *clo = data;
current->flags |= PF_NOFREEZE;
set_user_nice(current, -15);
while (!kthread_should_stop()||!list_empty(&clo->clo_list))
{
int err;
err = wait_event_interruptible(clo->clo_event, !list_empty(&clo->clo_list) ||
kthread_should_stop());
if(unlikely(err))
{
DEBUGP(KERN_ERR "cloop thread activated on error!? Continuing.\n");
continue;
}
if(!list_empty(&clo->clo_list))
{
struct request *req;
unsigned long flags;
int uptodate;
spin_lock_irq(&clo->queue_lock);
req = list_entry(clo->clo_list.next, struct request, queuelist);
list_del_init(&req->queuelist);
spin_unlock_irq(&clo->queue_lock);
uptodate = cloop_handle_request(clo, req);
spin_lock_irqsave(&clo->queue_lock, flags);
__blk_end_request_all(req, uptodate ? 0 : -EIO);
spin_unlock_irqrestore(&clo->queue_lock, flags);
}
}
DEBUGP(KERN_ERR "cloop_thread exited.\n");
return 0;
}
/* This is called by the kernel block queue management every now and then,
* with successive read requests qeued and sorted in a (hopefully)
* "most efficient way". spin_lock_irq() is being held by the kernel. */
static void cloop_do_request(struct request_queue *q)
{
struct request *req;
while((req = blk_fetch_request(q)) != NULL)
{
struct cloop_device *clo;
int rw;
/* quick sanity checks */
/* blk_fs_request() was removed in 2.6.36 */
if (unlikely(req == NULL || (req->cmd_type != REQ_TYPE_FS)))
goto error_continue;
rw = rq_data_dir(req);
if (unlikely(rw != READ && rw != READA))
{
DEBUGP("cloop_do_request: bad command\n");
goto error_continue;
}
clo = req->rq_disk->private_data;
if (unlikely(!clo->backing_file && !clo->suspended))
{
DEBUGP("cloop_do_request: not connected to a file\n");
goto error_continue;
}
list_add_tail(&req->queuelist, &clo->clo_list); /* Add to working list for thread */
wake_up(&clo->clo_event); /* Wake up cloop_thread */
continue; /* next request */
error_continue:
DEBUGP(KERN_ERR "cloop_do_request: Discarding request %p.\n", req);
req->errors++;
__blk_end_request_all(req, -EIO);
}
}
/* Read header and offsets from already opened file */
static int cloop_set_file(int cloop_num, struct file *file, char *filename)
{
struct cloop_device *clo = cloop_dev[cloop_num];
struct inode *inode;
char *bbuf=NULL;
unsigned int i, offsets_read, total_offsets;
int isblkdev;
int error = 0;
inode = file->f_dentry->d_inode;
isblkdev=S_ISBLK(inode->i_mode)?1:0;
if(!isblkdev&&!S_ISREG(inode->i_mode))
{
printk(KERN_ERR "%s: %s not a regular file or block device\n",
cloop_name, filename);
error=-EBADF; goto error_release;
}
clo->backing_file = file;
clo->backing_inode= inode ;
if(!isblkdev&&inode->i_size<sizeof(struct cloop_head))
{
printk(KERN_ERR "%s: %lu bytes (must be >= %u bytes)\n",
cloop_name, (unsigned long)inode->i_size,
(unsigned)sizeof(struct cloop_head));
error=-EBADF; goto error_release;
}
/* In suspended mode, we have done all checks necessary - FF */
if (clo->suspended)
return error;
if(isblkdev)
{
struct request_queue *q = bdev_get_queue(inode->i_bdev);
blk_queue_max_hw_sectors(clo->clo_queue, queue_max_hw_sectors(q)); /* Renamed in 2.6.34 */
blk_queue_max_segments(clo->clo_queue, queue_max_segments(q)); /* Renamed in 2.6.34 */
/* blk_queue_max_hw_segments(clo->clo_queue, queue_max_hw_segments(q)); */ /* Removed in 2.6.34 */
blk_queue_max_segment_size(clo->clo_queue, queue_max_segment_size(q));
blk_queue_segment_boundary(clo->clo_queue, queue_segment_boundary(q));
blk_queue_merge_bvec(clo->clo_queue, q->merge_bvec_fn);
clo->underlying_blksize = block_size(inode->i_bdev);
}
else
clo->underlying_blksize = PAGE_SIZE;
DEBUGP("Underlying blocksize is %u\n", clo->underlying_blksize);
bbuf = cloop_malloc(clo->underlying_blksize);
if(!bbuf)
{
printk(KERN_ERR "%s: out of kernel mem for block buffer (%lu bytes)\n",
cloop_name, (unsigned long)clo->underlying_blksize);
error=-ENOMEM; goto error_release;
}
total_offsets = 1; /* Dummy total_offsets: will be filled in first time around */
for (i = 0, offsets_read = 0; offsets_read < total_offsets; i++)
{
unsigned int offset = 0, num_readable;
size_t bytes_read = cloop_read_from_file(clo, file, bbuf,
i*clo->underlying_blksize,
clo->underlying_blksize);
if(bytes_read != clo->underlying_blksize)
{
printk(KERN_ERR "%s: Bad file, read() of first %lu bytes returned %d.\n",
cloop_name, (unsigned long)clo->underlying_blksize, (int)bytes_read);
error=-EBADF;
goto error_release;
}
/* Header will be in block zero */
if(i==0)
{
memcpy(&clo->head, bbuf, sizeof(struct cloop_head));
offset = sizeof(struct cloop_head);
if (ntohl(clo->head.block_size) % 512 != 0)
{
printk(KERN_ERR "%s: blocksize %u not multiple of 512\n",
cloop_name, ntohl(clo->head.block_size));
error=-EBADF; goto error_release;
}
if (clo->head.preamble[0x0B]!='V'||clo->head.preamble[0x0C]<'1')
{
printk(KERN_ERR "%s: Cannot read old 32-bit (version 0.68) images, "
"please use an older version of %s for this file.\n",
cloop_name, cloop_name);
error=-EBADF; goto error_release;
}
if (clo->head.preamble[0x0C]<'2')
{
printk(KERN_ERR "%s: Cannot read old architecture-dependent "
"(format <= 1.0) images, please use an older "
"version of %s for this file.\n",
cloop_name, cloop_name);
error=-EBADF; goto error_release;
}
total_offsets=ntohl(clo->head.num_blocks)+1;
if (!isblkdev && (sizeof(struct cloop_head)+sizeof(loff_t)*
total_offsets > inode->i_size))
{
printk(KERN_ERR "%s: file too small for %u blocks\n",
cloop_name, ntohl(clo->head.num_blocks));
error=-EBADF; goto error_release;
}
clo->offsets = cloop_malloc(sizeof(loff_t) * total_offsets);
if (!clo->offsets)
{
printk(KERN_ERR "%s: out of kernel mem for offsets\n", cloop_name);
error=-ENOMEM; goto error_release;
}
}
num_readable = MIN(total_offsets - offsets_read,
(clo->underlying_blksize - offset)
/ sizeof(loff_t));
memcpy(&clo->offsets[offsets_read], bbuf+offset, num_readable * sizeof(loff_t));
offsets_read += num_readable;
}
{ /* Search for largest block rather than estimate. KK. */
int i;
for(i=0;i<total_offsets-1;i++)
{
loff_t d=be64_to_cpu(clo->offsets[i+1]) - be64_to_cpu(clo->offsets[i]);
clo->largest_block=MAX(clo->largest_block,d);
}
printk(KERN_INFO "%s: %s: %u blocks, %u bytes/block, largest block is %lu bytes.\n",
cloop_name, filename, ntohl(clo->head.num_blocks),
ntohl(clo->head.block_size), clo->largest_block);
}
/* Combo kmalloc used too large chunks (>130000). */
{
int i;
for(i=0;i<BUFFERED_BLOCKS;i++)
{
clo->buffer[i] = cloop_malloc(ntohl(clo->head.block_size));
if(!clo->buffer[i])
{
printk(KERN_ERR "%s: out of memory for buffer %lu\n",
cloop_name, (unsigned long) ntohl(clo->head.block_size));
error=-ENOMEM; goto error_release_free;
}
}
}
clo->compressed_buffer = cloop_malloc(clo->largest_block);
if(!clo->compressed_buffer)
{
printk(KERN_ERR "%s: out of memory for compressed buffer %lu\n",
cloop_name, clo->largest_block);
error=-ENOMEM; goto error_release_free_buffer;
}
clo->zstream.workspace = cloop_malloc(zlib_inflate_workspacesize());
if(!clo->zstream.workspace)
{
printk(KERN_ERR "%s: out of mem for zlib working area %u\n",
cloop_name, zlib_inflate_workspacesize());
error=-ENOMEM; goto error_release_free_all;
}
zlib_inflateInit(&clo->zstream);
if(!isblkdev &&
be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]) != inode->i_size)
{
printk(KERN_ERR "%s: final offset wrong (%Lu not %Lu)\n",
cloop_name,
be64_to_cpu(clo->offsets[ntohl(clo->head.num_blocks)]),
inode->i_size);
cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace=NULL;
goto error_release_free_all;
}
{
int i;
for(i=0; i<BUFFERED_BLOCKS; i++) clo->buffered_blocknum[i] = -1;
clo->current_bufnum=0;
}
set_capacity(clo->clo_disk, (sector_t)(ntohl(clo->head.num_blocks)*
(ntohl(clo->head.block_size)>>9)));
clo->clo_thread = kthread_create(cloop_thread, clo, "cloop%d", cloop_num);
if(IS_ERR(clo->clo_thread))
{
error = PTR_ERR(clo->clo_thread);
clo->clo_thread=NULL;
goto error_release_free_all;
}
if(preload > 0)
{
clo->preload_array_size = ((preload<=ntohl(clo->head.num_blocks))?preload:ntohl(clo->head.num_blocks));
clo->preload_size = 0;
if((clo->preload_cache = cloop_malloc(clo->preload_array_size * sizeof(char *))) != NULL)
{
int i;
for(i=0; i<clo->preload_array_size; i++)
{
if((clo->preload_cache[i] = cloop_malloc(ntohl(clo->head.block_size))) == NULL)
{ /* Out of memory */
printk(KERN_WARNING "%s: cloop_malloc(%d) failed for preload_cache[%d] (ignored).\n",
cloop_name, ntohl(clo->head.block_size), i);
break;
}
}
clo->preload_size = i;
for(i=0; i<clo->preload_size; i++)
{
int buffered_blocknum = cloop_load_buffer(clo,i);
if(buffered_blocknum >= 0)
{
memcpy(clo->preload_cache[i], clo->buffer[buffered_blocknum],
ntohl(clo->head.block_size));
}
else
{
printk(KERN_WARNING "%s: can't read block %d into preload cache, set to zero.\n",
cloop_name, i);
memset(clo->preload_cache[i], 0, ntohl(clo->head.block_size));
}
}
printk(KERN_INFO "%s: preloaded %d blocks into cache.\n", cloop_name,
(int)clo->preload_size);
}
else
{
/* It is not a fatal error if cloop_malloc(clo->preload_size)
* fails, then we just go without cache, but we should at least
* let the user know. */
printk(KERN_WARNING "%s: cloop_malloc(%d) failed, continuing without preloaded buffers.\n",
cloop_name, (int)(clo->preload_size * sizeof(char *)));
clo->preload_array_size = clo->preload_size = 0;
}
}
wake_up_process(clo->clo_thread);
/* Uncheck */
return error;
error_release_free_all:
cloop_free(clo->compressed_buffer, clo->largest_block);
clo->compressed_buffer=NULL;
error_release_free_buffer:
{
int i;
for(i=0; i<BUFFERED_BLOCKS; i++)
{
if(clo->buffer[i])
{
cloop_free(clo->buffer[i], ntohl(clo->head.block_size));
clo->buffer[i]=NULL;
}
}
}
error_release_free:
cloop_free(clo->offsets, sizeof(loff_t) * total_offsets);
clo->offsets=NULL;
error_release:
if(bbuf) cloop_free(bbuf, clo->underlying_blksize);
clo->backing_file=NULL;
return error;
}
/* Get file from ioctl arg (only losetup) */
static int cloop_set_fd(int cloop_num, struct file *clo_file,
struct block_device *bdev, unsigned int arg)
{
struct cloop_device *clo = cloop_dev[cloop_num];
struct file *file=NULL;
int error = 0;
/* Already an allocated file present */
if(clo->backing_file) return -EBUSY;
file = fget(arg); /* get filp struct from ioctl arg fd */
if(!file) return -EBADF;
error=cloop_set_file(cloop_num,file,"losetup_file");
set_device_ro(bdev, 1);
if(error) fput(file);
return error;
}
/* Drop file and free buffers, both ioctl and initial_file */
static int cloop_clr_fd(int cloop_num, struct block_device *bdev)
{
struct cloop_device *clo = cloop_dev[cloop_num];
struct file *filp = clo->backing_file;
int i;
if(clo->refcnt > 1) /* we needed one fd for the ioctl */
return -EBUSY;
if(filp==NULL) return -EINVAL;
if(clo->clo_thread) { kthread_stop(clo->clo_thread); clo->clo_thread=NULL; }
if(filp!=initial_file) fput(filp);
else { filp_close(initial_file,0); initial_file=NULL; }
clo->backing_file = NULL;
clo->backing_inode = NULL;
if(clo->offsets) { cloop_free(clo->offsets, clo->underlying_blksize); clo->offsets = NULL; }
if(clo->preload_cache)
{
for(i=0; i < clo->preload_size; i++)
cloop_free(clo->preload_cache[i], ntohl(clo->head.block_size));
cloop_free(clo->preload_cache, clo->preload_array_size * sizeof(char *));
clo->preload_cache = NULL;
clo->preload_size = clo->preload_array_size = 0;
}
for(i=0; i<BUFFERED_BLOCKS; i++)
if(clo->buffer[i]) { cloop_free(clo->buffer[i], ntohl(clo->head.block_size)); clo->buffer[i]=NULL; }
if(clo->compressed_buffer) { cloop_free(clo->compressed_buffer, clo->largest_block); clo->compressed_buffer = NULL; }
zlib_inflateEnd(&clo->zstream);
if(clo->zstream.workspace) { cloop_free(clo->zstream.workspace, zlib_inflate_workspacesize()); clo->zstream.workspace = NULL; }
if(bdev) invalidate_bdev(bdev);
if(clo->clo_disk) set_capacity(clo->clo_disk, 0);
return 0;
}
static int clo_suspend_fd(int cloop_num)
{
struct cloop_device *clo = cloop_dev[cloop_num];
struct file *filp = clo->backing_file;
if(filp==NULL || clo->suspended) return -EINVAL;
/* Suspend all running requests - FF */
clo->suspended=1;
if(filp!=initial_file) fput(filp);
else { filp_close(initial_file,0); initial_file=NULL; }
clo->backing_file = NULL;
clo->backing_inode = NULL;
return 0;
}
/* Copied from loop.c, stripped down to the really necessary */
static int cloop_set_status(struct cloop_device *clo,
const struct loop_info64 *info)
{
if (!clo->backing_file) return -ENXIO;
memcpy(clo->clo_file_name, info->lo_file_name, LO_NAME_SIZE);
clo->clo_file_name[LO_NAME_SIZE-1] = 0;
return 0;
}
static int cloop_get_status(struct cloop_device *clo,
struct loop_info64 *info)
{
struct file *file = clo->backing_file;
struct kstat stat;
int err;
if (!file) return -ENXIO;
err = vfs_getattr(&file->f_path, &stat);
if (err) return err;
memset(info, 0, sizeof(*info));
info->lo_number = clo->clo_number;
info->lo_device = huge_encode_dev(stat.dev);
info->lo_inode = stat.ino;
info->lo_rdevice = huge_encode_dev(clo->isblkdev ? stat.rdev : stat.dev);
info->lo_offset = 0;
info->lo_sizelimit = 0;
info->lo_flags = 0;
memcpy(info->lo_file_name, clo->clo_file_name, LO_NAME_SIZE);
return 0;
}
static void cloop_info64_from_old(const struct loop_info *info,
struct loop_info64 *info64)
{
memset(info64, 0, sizeof(*info64));
info64->lo_number = info->lo_number;
info64->lo_device = info->lo_device;
info64->lo_inode = info->lo_inode;
info64->lo_rdevice = info->lo_rdevice;
info64->lo_offset = info->lo_offset;
info64->lo_sizelimit = 0;
info64->lo_flags = info->lo_flags;
info64->lo_init[0] = info->lo_init[0];
info64->lo_init[1] = info->lo_init[1];
memcpy(info64->lo_file_name, info->lo_name, LO_NAME_SIZE);
}
static int cloop_info64_to_old(const struct loop_info64 *info64,
struct loop_info *info)
{
memset(info, 0, sizeof(*info));
info->lo_number = info64->lo_number;
info->lo_device = info64->lo_device;
info->lo_inode = info64->lo_inode;
info->lo_rdevice = info64->lo_rdevice;
info->lo_offset = info64->lo_offset;
info->lo_flags = info64->lo_flags;
info->lo_init[0] = info64->lo_init[0];
info->lo_init[1] = info64->lo_init[1];
memcpy(info->lo_name, info64->lo_file_name, LO_NAME_SIZE);
return 0;
}
static int cloop_set_status_old(struct cloop_device *clo,
const struct loop_info __user *arg)
{
struct loop_info info;
struct loop_info64 info64;
if (copy_from_user(&info, arg, sizeof (struct loop_info))) return -EFAULT;
cloop_info64_from_old(&info, &info64);
return cloop_set_status(clo, &info64);
}
static int cloop_set_status64(struct cloop_device *clo,
const struct loop_info64 __user *arg)
{
struct loop_info64 info64;
if (copy_from_user(&info64, arg, sizeof (struct loop_info64)))
return -EFAULT;
return cloop_set_status(clo, &info64);
}
static int cloop_get_status_old(struct cloop_device *clo,
struct loop_info __user *arg)
{
struct loop_info info;
struct loop_info64 info64;
int err = 0;
if (!arg) err = -EINVAL;
if (!err) err = cloop_get_status(clo, &info64);
if (!err) err = cloop_info64_to_old(&info64, &info);
if (!err && copy_to_user(arg, &info, sizeof(info))) err = -EFAULT;
return err;
}
static int cloop_get_status64(struct cloop_device *clo,
struct loop_info64 __user *arg)
{
struct loop_info64 info64;
int err = 0;
if (!arg) err = -EINVAL;
if (!err) err = cloop_get_status(clo, &info64);
if (!err && copy_to_user(arg, &info64, sizeof(info64))) err = -EFAULT;
return err;
}
/* EOF get/set_status */
static int cloop_ioctl(struct block_device *bdev, fmode_t mode,
unsigned int cmd, unsigned long arg)
{
struct cloop_device *clo;
int cloop_num, err=0;
if (!bdev) return -EINVAL;
cloop_num = MINOR(bdev->bd_dev);
if (cloop_num < 0 || cloop_num > cloop_count-1) return -ENODEV;
clo = cloop_dev[cloop_num];
mutex_lock(&clo->clo_ctl_mutex);
switch (cmd)
{ /* We use the same ioctls that loop does */
case LOOP_CHANGE_FD:
case LOOP_SET_FD:
err = cloop_set_fd(cloop_num, NULL, bdev, arg);
if (err == 0 && clo->suspended)
{
/* Okay, we have again a backing file - get reqs again - FF */
clo->suspended=0;
}
break;
case LOOP_CLR_FD:
err = cloop_clr_fd(cloop_num, bdev);
break;
case LOOP_SET_STATUS:
err = cloop_set_status_old(clo, (struct loop_info __user *) arg);
break;
case LOOP_GET_STATUS:
err = cloop_get_status_old(clo, (struct loop_info __user *) arg);
break;
case LOOP_SET_STATUS64:
err = cloop_set_status64(clo, (struct loop_info64 __user *) arg);
break;
case LOOP_GET_STATUS64:
err = cloop_get_status64(clo, (struct loop_info64 __user *) arg);
break;
case CLOOP_SUSPEND:
err = clo_suspend_fd(cloop_num);
break;
default:
err = -EINVAL;
}
mutex_unlock(&clo->clo_ctl_mutex);
return err;
}
#ifdef CONFIG_COMPAT
static int cloop_compat_ioctl(struct block_device *bdev, fmode_t mode,
unsigned int cmd, unsigned long arg)
{
switch(cmd) {
case LOOP_SET_CAPACITY: /* Change arg */
case LOOP_CLR_FD: /* Change arg */
case LOOP_GET_STATUS64: /* Change arg */
case LOOP_SET_STATUS64: /* Change arg */
arg = (unsigned long) compat_ptr(arg);
case LOOP_SET_STATUS: /* unchanged */
case LOOP_GET_STATUS: /* unchanged */
case LOOP_SET_FD: /* unchanged */
case LOOP_CHANGE_FD: /* unchanged */
return cloop_ioctl(bdev, mode, cmd, arg);
break;
}
return -ENOIOCTLCMD;
}
#endif
static int cloop_open(struct block_device *bdev, fmode_t mode)
{
int cloop_num;
if(!bdev) return -EINVAL;
cloop_num=MINOR(bdev->bd_dev);
if(cloop_num > cloop_count-1) return -ENODEV;
/* Allow write open for ioctl, but not for mount. */
/* losetup uses write-open and flags=0x8002 to set a new file */
if(mode & FMODE_WRITE)
{
printk(KERN_WARNING "%s: Can't open device read-write in mode 0x%x\n", cloop_name, mode);
return -EROFS;
}
cloop_dev[cloop_num]->refcnt+=1;
return 0;
}
static void cloop_close(struct gendisk *disk, fmode_t mode)
{
int cloop_num;
if(!disk) return;
cloop_num=((struct cloop_device *)disk->private_data)->clo_number;
if(cloop_num < 0 || cloop_num > (cloop_count-1)) return;
cloop_dev[cloop_num]->refcnt-=1;
}
static struct block_device_operations clo_fops =
{
owner: THIS_MODULE,
open: cloop_open,
release: cloop_close,
#ifdef CONFIG_COMPAT
compat_ioctl: cloop_compat_ioctl,
#endif
ioctl: cloop_ioctl
/* locked_ioctl ceased to exist in 2.6.36 */
};
static int cloop_register_blkdev(int major_nr)
{
return register_blkdev(major_nr, cloop_name);
}
static int cloop_unregister_blkdev(void)
{
unregister_blkdev(cloop_major, cloop_name);
return 0;
}
static int cloop_alloc(int cloop_num)
{
struct cloop_device *clo = (struct cloop_device *) cloop_malloc(sizeof(struct cloop_device));;
if(clo == NULL) goto error_out;
cloop_dev[cloop_num] = clo;
memset(clo, 0, sizeof(struct cloop_device));
clo->clo_number = cloop_num;
clo->clo_thread = NULL;
init_waitqueue_head(&clo->clo_event);
spin_lock_init(&clo->queue_lock);
mutex_init(&clo->clo_ctl_mutex);
INIT_LIST_HEAD(&clo->clo_list);
clo->clo_queue = blk_init_queue(cloop_do_request, &clo->queue_lock);
if(!clo->clo_queue)
{
printk(KERN_ERR "%s: Unable to alloc queue[%d]\n", cloop_name, cloop_num);
goto error_out;
}
clo->clo_queue->queuedata = clo;
clo->clo_disk = alloc_disk(1);
if(!clo->clo_disk)
{
printk(KERN_ERR "%s: Unable to alloc disk[%d]\n", cloop_name, cloop_num);
goto error_disk;
}
clo->clo_disk->major = cloop_major;
clo->clo_disk->first_minor = cloop_num;
clo->clo_disk->fops = &clo_fops;
clo->clo_disk->queue = clo->clo_queue;
clo->clo_disk->private_data = clo;
sprintf(clo->clo_disk->disk_name, "%s%d", cloop_name, cloop_num);
add_disk(clo->clo_disk);
return 0;
error_disk:
blk_cleanup_queue(clo->clo_queue);
error_out:
return -ENOMEM;
}
static void cloop_dealloc(int cloop_num)
{