1 #ifndef _IDE_H 2 #define _IDE_H 3 /* 4 * linux/include/linux/ide.h 5 * 6 * Copyright (C) 1994-2002 Linus Torvalds & authors 7 */ 8 9 #include <linux/init.h> 10 #include <linux/ioport.h> 11 #include <linux/ata.h> 12 #include <linux/blkdev.h> 13 #include <linux/proc_fs.h> 14 #include <linux/interrupt.h> 15 #include <linux/bitops.h> 16 #include <linux/bio.h> 17 #include <linux/pci.h> 18 #include <linux/completion.h> 19 #include <linux/pm.h> 20 #include <linux/mutex.h> 21 /* for request_sense */ 22 #include <linux/cdrom.h> 23 #include <scsi/scsi_cmnd.h> 24 #include <asm/byteorder.h> 25 #include <asm/io.h> 26 27 #if defined(CONFIG_CRIS) || defined(CONFIG_FRV) || defined(CONFIG_MN10300) 28 # define SUPPORT_VLB_SYNC 0 29 #else 30 # define SUPPORT_VLB_SYNC 1 31 #endif 32 33 /* 34 * Probably not wise to fiddle with these 35 */ 36 #define IDE_DEFAULT_MAX_FAILURES 1 37 #define ERROR_MAX 8 /* Max read/write errors per sector */ 38 #define ERROR_RESET 3 /* Reset controller every 4th retry */ 39 #define ERROR_RECAL 1 /* Recalibrate every 2nd retry */ 40 41 struct device; 42 43 /* values for ide_request.type */ 44 enum ata_priv_type { 45 ATA_PRIV_MISC, 46 ATA_PRIV_TASKFILE, 47 ATA_PRIV_PC, 48 ATA_PRIV_SENSE, /* sense request */ 49 ATA_PRIV_PM_SUSPEND, /* suspend request */ 50 ATA_PRIV_PM_RESUME, /* resume request */ 51 }; 52 53 struct ide_request { 54 struct scsi_request sreq; 55 u8 sense[SCSI_SENSE_BUFFERSIZE]; 56 u8 type; 57 }; 58 59 static inline struct ide_request *ide_req(struct request *rq) 60 { 61 return blk_mq_rq_to_pdu(rq); 62 } 63 64 static inline bool ata_misc_request(struct request *rq) 65 { 66 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_MISC; 67 } 68 69 static inline bool ata_taskfile_request(struct request *rq) 70 { 71 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_TASKFILE; 72 } 73 74 static inline bool ata_pc_request(struct request *rq) 75 { 76 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_PC; 77 } 78 79 static inline bool ata_sense_request(struct request *rq) 80 { 81 return blk_rq_is_private(rq) && ide_req(rq)->type == ATA_PRIV_SENSE; 82 } 83 84 static inline bool ata_pm_request(struct request *rq) 85 { 86 return blk_rq_is_private(rq) && 87 (ide_req(rq)->type == ATA_PRIV_PM_SUSPEND || 88 ide_req(rq)->type == ATA_PRIV_PM_RESUME); 89 } 90 91 /* Error codes returned in result to the higher part of the driver. */ 92 enum { 93 IDE_DRV_ERROR_GENERAL = 101, 94 IDE_DRV_ERROR_FILEMARK = 102, 95 IDE_DRV_ERROR_EOD = 103, 96 }; 97 98 /* 99 * Definitions for accessing IDE controller registers 100 */ 101 #define IDE_NR_PORTS (10) 102 103 struct ide_io_ports { 104 unsigned long data_addr; 105 106 union { 107 unsigned long error_addr; /* read: error */ 108 unsigned long feature_addr; /* write: feature */ 109 }; 110 111 unsigned long nsect_addr; 112 unsigned long lbal_addr; 113 unsigned long lbam_addr; 114 unsigned long lbah_addr; 115 116 unsigned long device_addr; 117 118 union { 119 unsigned long status_addr; /* read: status */ 120 unsigned long command_addr; /* write: command */ 121 }; 122 123 unsigned long ctl_addr; 124 125 unsigned long irq_addr; 126 }; 127 128 #define OK_STAT(stat,good,bad) (((stat)&((good)|(bad)))==(good)) 129 130 #define BAD_R_STAT (ATA_BUSY | ATA_ERR) 131 #define BAD_W_STAT (BAD_R_STAT | ATA_DF) 132 #define BAD_STAT (BAD_R_STAT | ATA_DRQ) 133 #define DRIVE_READY (ATA_DRDY | ATA_DSC) 134 135 #define BAD_CRC (ATA_ABORTED | ATA_ICRC) 136 137 #define SATA_NR_PORTS (3) /* 16 possible ?? */ 138 139 #define SATA_STATUS_OFFSET (0) 140 #define SATA_ERROR_OFFSET (1) 141 #define SATA_CONTROL_OFFSET (2) 142 143 /* 144 * Our Physical Region Descriptor (PRD) table should be large enough 145 * to handle the biggest I/O request we are likely to see. Since requests 146 * can have no more than 256 sectors, and since the typical blocksize is 147 * two or more sectors, we could get by with a limit of 128 entries here for 148 * the usual worst case. Most requests seem to include some contiguous blocks, 149 * further reducing the number of table entries required. 150 * 151 * The driver reverts to PIO mode for individual requests that exceed 152 * this limit (possible with 512 byte blocksizes, eg. MSDOS f/s), so handling 153 * 100% of all crazy scenarios here is not necessary. 154 * 155 * As it turns out though, we must allocate a full 4KB page for this, 156 * so the two PRD tables (ide0 & ide1) will each get half of that, 157 * allowing each to have about 256 entries (8 bytes each) from this. 158 */ 159 #define PRD_BYTES 8 160 #define PRD_ENTRIES 256 161 162 /* 163 * Some more useful definitions 164 */ 165 #define PARTN_BITS 6 /* number of minor dev bits for partitions */ 166 #define MAX_DRIVES 2 /* per interface; 2 assumed by lots of code */ 167 #define SECTOR_SIZE 512 168 169 /* 170 * Timeouts for various operations: 171 */ 172 enum { 173 /* spec allows up to 20ms, but CF cards and SSD drives need more */ 174 WAIT_DRQ = 1 * HZ, /* 1s */ 175 /* some laptops are very slow */ 176 WAIT_READY = 5 * HZ, /* 5s */ 177 /* should be less than 3ms (?), if all ATAPI CD is closed at boot */ 178 WAIT_PIDENTIFY = 10 * HZ, /* 10s */ 179 /* worst case when spinning up */ 180 WAIT_WORSTCASE = 30 * HZ, /* 30s */ 181 /* maximum wait for an IRQ to happen */ 182 WAIT_CMD = 10 * HZ, /* 10s */ 183 /* Some drives require a longer IRQ timeout. */ 184 WAIT_FLOPPY_CMD = 50 * HZ, /* 50s */ 185 /* 186 * Some drives (for example, Seagate STT3401A Travan) require a very 187 * long timeout, because they don't return an interrupt or clear their 188 * BSY bit until after the command completes (even retension commands). 189 */ 190 WAIT_TAPE_CMD = 900 * HZ, /* 900s */ 191 /* minimum sleep time */ 192 WAIT_MIN_SLEEP = HZ / 50, /* 20ms */ 193 }; 194 195 /* 196 * Op codes for special requests to be handled by ide_special_rq(). 197 * Values should be in the range of 0x20 to 0x3f. 198 */ 199 #define REQ_DRIVE_RESET 0x20 200 #define REQ_DEVSET_EXEC 0x21 201 #define REQ_PARK_HEADS 0x22 202 #define REQ_UNPARK_HEADS 0x23 203 204 /* 205 * hwif_chipset_t is used to keep track of the specific hardware 206 * chipset used by each IDE interface, if known. 207 */ 208 enum { ide_unknown, ide_generic, ide_pci, 209 ide_cmd640, ide_dtc2278, ide_ali14xx, 210 ide_qd65xx, ide_umc8672, ide_ht6560b, 211 ide_4drives, ide_pmac, ide_acorn, 212 ide_au1xxx, ide_palm3710 213 }; 214 215 typedef u8 hwif_chipset_t; 216 217 /* 218 * Structure to hold all information about the location of this port 219 */ 220 struct ide_hw { 221 union { 222 struct ide_io_ports io_ports; 223 unsigned long io_ports_array[IDE_NR_PORTS]; 224 }; 225 226 int irq; /* our irq number */ 227 struct device *dev, *parent; 228 unsigned long config; 229 }; 230 231 static inline void ide_std_init_ports(struct ide_hw *hw, 232 unsigned long io_addr, 233 unsigned long ctl_addr) 234 { 235 unsigned int i; 236 237 for (i = 0; i <= 7; i++) 238 hw->io_ports_array[i] = io_addr++; 239 240 hw->io_ports.ctl_addr = ctl_addr; 241 } 242 243 #define MAX_HWIFS 10 244 245 /* 246 * Now for the data we need to maintain per-drive: ide_drive_t 247 */ 248 249 #define ide_scsi 0x21 250 #define ide_disk 0x20 251 #define ide_optical 0x7 252 #define ide_cdrom 0x5 253 #define ide_tape 0x1 254 #define ide_floppy 0x0 255 256 /* 257 * Special Driver Flags 258 */ 259 enum { 260 IDE_SFLAG_SET_GEOMETRY = (1 << 0), 261 IDE_SFLAG_RECALIBRATE = (1 << 1), 262 IDE_SFLAG_SET_MULTMODE = (1 << 2), 263 }; 264 265 /* 266 * Status returned from various ide_ functions 267 */ 268 typedef enum { 269 ide_stopped, /* no drive operation was started */ 270 ide_started, /* a drive operation was started, handler was set */ 271 } ide_startstop_t; 272 273 enum { 274 IDE_VALID_ERROR = (1 << 1), 275 IDE_VALID_FEATURE = IDE_VALID_ERROR, 276 IDE_VALID_NSECT = (1 << 2), 277 IDE_VALID_LBAL = (1 << 3), 278 IDE_VALID_LBAM = (1 << 4), 279 IDE_VALID_LBAH = (1 << 5), 280 IDE_VALID_DEVICE = (1 << 6), 281 IDE_VALID_LBA = IDE_VALID_LBAL | 282 IDE_VALID_LBAM | 283 IDE_VALID_LBAH, 284 IDE_VALID_OUT_TF = IDE_VALID_FEATURE | 285 IDE_VALID_NSECT | 286 IDE_VALID_LBA, 287 IDE_VALID_IN_TF = IDE_VALID_NSECT | 288 IDE_VALID_LBA, 289 IDE_VALID_OUT_HOB = IDE_VALID_OUT_TF, 290 IDE_VALID_IN_HOB = IDE_VALID_ERROR | 291 IDE_VALID_NSECT | 292 IDE_VALID_LBA, 293 }; 294 295 enum { 296 IDE_TFLAG_LBA48 = (1 << 0), 297 IDE_TFLAG_WRITE = (1 << 1), 298 IDE_TFLAG_CUSTOM_HANDLER = (1 << 2), 299 IDE_TFLAG_DMA_PIO_FALLBACK = (1 << 3), 300 /* force 16-bit I/O operations */ 301 IDE_TFLAG_IO_16BIT = (1 << 4), 302 /* struct ide_cmd was allocated using kmalloc() */ 303 IDE_TFLAG_DYN = (1 << 5), 304 IDE_TFLAG_FS = (1 << 6), 305 IDE_TFLAG_MULTI_PIO = (1 << 7), 306 IDE_TFLAG_SET_XFER = (1 << 8), 307 }; 308 309 enum { 310 IDE_FTFLAG_FLAGGED = (1 << 0), 311 IDE_FTFLAG_SET_IN_FLAGS = (1 << 1), 312 IDE_FTFLAG_OUT_DATA = (1 << 2), 313 IDE_FTFLAG_IN_DATA = (1 << 3), 314 }; 315 316 struct ide_taskfile { 317 u8 data; /* 0: data byte (for TASKFILE ioctl) */ 318 union { /* 1: */ 319 u8 error; /* read: error */ 320 u8 feature; /* write: feature */ 321 }; 322 u8 nsect; /* 2: number of sectors */ 323 u8 lbal; /* 3: LBA low */ 324 u8 lbam; /* 4: LBA mid */ 325 u8 lbah; /* 5: LBA high */ 326 u8 device; /* 6: device select */ 327 union { /* 7: */ 328 u8 status; /* read: status */ 329 u8 command; /* write: command */ 330 }; 331 }; 332 333 struct ide_cmd { 334 struct ide_taskfile tf; 335 struct ide_taskfile hob; 336 struct { 337 struct { 338 u8 tf; 339 u8 hob; 340 } out, in; 341 } valid; 342 343 u16 tf_flags; 344 u8 ftf_flags; /* for TASKFILE ioctl */ 345 int protocol; 346 347 int sg_nents; /* number of sg entries */ 348 int orig_sg_nents; 349 int sg_dma_direction; /* DMA transfer direction */ 350 351 unsigned int nbytes; 352 unsigned int nleft; 353 unsigned int last_xfer_len; 354 355 struct scatterlist *cursg; 356 unsigned int cursg_ofs; 357 358 struct request *rq; /* copy of request */ 359 }; 360 361 /* ATAPI packet command flags */ 362 enum { 363 /* set when an error is considered normal - no retry (ide-tape) */ 364 PC_FLAG_ABORT = (1 << 0), 365 PC_FLAG_SUPPRESS_ERROR = (1 << 1), 366 PC_FLAG_WAIT_FOR_DSC = (1 << 2), 367 PC_FLAG_DMA_OK = (1 << 3), 368 PC_FLAG_DMA_IN_PROGRESS = (1 << 4), 369 PC_FLAG_DMA_ERROR = (1 << 5), 370 PC_FLAG_WRITING = (1 << 6), 371 }; 372 373 #define ATAPI_WAIT_PC (60 * HZ) 374 375 struct ide_atapi_pc { 376 /* actual packet bytes */ 377 u8 c[12]; 378 /* incremented on each retry */ 379 int retries; 380 int error; 381 382 /* bytes to transfer */ 383 int req_xfer; 384 385 /* the corresponding request */ 386 struct request *rq; 387 388 unsigned long flags; 389 390 /* 391 * those are more or less driver-specific and some of them are subject 392 * to change/removal later. 393 */ 394 unsigned long timeout; 395 }; 396 397 struct ide_devset; 398 struct ide_driver; 399 400 #ifdef CONFIG_BLK_DEV_IDEACPI 401 struct ide_acpi_drive_link; 402 struct ide_acpi_hwif_link; 403 #endif 404 405 struct ide_drive_s; 406 407 struct ide_disk_ops { 408 int (*check)(struct ide_drive_s *, const char *); 409 int (*get_capacity)(struct ide_drive_s *); 410 void (*unlock_native_capacity)(struct ide_drive_s *); 411 void (*setup)(struct ide_drive_s *); 412 void (*flush)(struct ide_drive_s *); 413 int (*init_media)(struct ide_drive_s *, struct gendisk *); 414 int (*set_doorlock)(struct ide_drive_s *, struct gendisk *, 415 int); 416 ide_startstop_t (*do_request)(struct ide_drive_s *, struct request *, 417 sector_t); 418 int (*ioctl)(struct ide_drive_s *, struct block_device *, 419 fmode_t, unsigned int, unsigned long); 420 }; 421 422 /* ATAPI device flags */ 423 enum { 424 IDE_AFLAG_DRQ_INTERRUPT = (1 << 0), 425 426 /* ide-cd */ 427 /* Drive cannot eject the disc. */ 428 IDE_AFLAG_NO_EJECT = (1 << 1), 429 /* Drive is a pre ATAPI 1.2 drive. */ 430 IDE_AFLAG_PRE_ATAPI12 = (1 << 2), 431 /* TOC addresses are in BCD. */ 432 IDE_AFLAG_TOCADDR_AS_BCD = (1 << 3), 433 /* TOC track numbers are in BCD. */ 434 IDE_AFLAG_TOCTRACKS_AS_BCD = (1 << 4), 435 /* Saved TOC information is current. */ 436 IDE_AFLAG_TOC_VALID = (1 << 6), 437 /* We think that the drive door is locked. */ 438 IDE_AFLAG_DOOR_LOCKED = (1 << 7), 439 /* SET_CD_SPEED command is unsupported. */ 440 IDE_AFLAG_NO_SPEED_SELECT = (1 << 8), 441 IDE_AFLAG_VERTOS_300_SSD = (1 << 9), 442 IDE_AFLAG_VERTOS_600_ESD = (1 << 10), 443 IDE_AFLAG_SANYO_3CD = (1 << 11), 444 IDE_AFLAG_FULL_CAPS_PAGE = (1 << 12), 445 IDE_AFLAG_PLAY_AUDIO_OK = (1 << 13), 446 IDE_AFLAG_LE_SPEED_FIELDS = (1 << 14), 447 448 /* ide-floppy */ 449 /* Avoid commands not supported in Clik drive */ 450 IDE_AFLAG_CLIK_DRIVE = (1 << 15), 451 /* Requires BH algorithm for packets */ 452 IDE_AFLAG_ZIP_DRIVE = (1 << 16), 453 /* Supports format progress report */ 454 IDE_AFLAG_SRFP = (1 << 17), 455 456 /* ide-tape */ 457 IDE_AFLAG_IGNORE_DSC = (1 << 18), 458 /* 0 When the tape position is unknown */ 459 IDE_AFLAG_ADDRESS_VALID = (1 << 19), 460 /* Device already opened */ 461 IDE_AFLAG_BUSY = (1 << 20), 462 /* Attempt to auto-detect the current user block size */ 463 IDE_AFLAG_DETECT_BS = (1 << 21), 464 /* Currently on a filemark */ 465 IDE_AFLAG_FILEMARK = (1 << 22), 466 /* 0 = no tape is loaded, so we don't rewind after ejecting */ 467 IDE_AFLAG_MEDIUM_PRESENT = (1 << 23), 468 469 IDE_AFLAG_NO_AUTOCLOSE = (1 << 24), 470 }; 471 472 /* device flags */ 473 enum { 474 /* restore settings after device reset */ 475 IDE_DFLAG_KEEP_SETTINGS = (1 << 0), 476 /* device is using DMA for read/write */ 477 IDE_DFLAG_USING_DMA = (1 << 1), 478 /* okay to unmask other IRQs */ 479 IDE_DFLAG_UNMASK = (1 << 2), 480 /* don't attempt flushes */ 481 IDE_DFLAG_NOFLUSH = (1 << 3), 482 /* DSC overlap */ 483 IDE_DFLAG_DSC_OVERLAP = (1 << 4), 484 /* give potential excess bandwidth */ 485 IDE_DFLAG_NICE1 = (1 << 5), 486 /* device is physically present */ 487 IDE_DFLAG_PRESENT = (1 << 6), 488 /* disable Host Protected Area */ 489 IDE_DFLAG_NOHPA = (1 << 7), 490 /* id read from device (synthetic if not set) */ 491 IDE_DFLAG_ID_READ = (1 << 8), 492 IDE_DFLAG_NOPROBE = (1 << 9), 493 /* need to do check_media_change() */ 494 IDE_DFLAG_REMOVABLE = (1 << 10), 495 /* needed for removable devices */ 496 IDE_DFLAG_ATTACH = (1 << 11), 497 IDE_DFLAG_FORCED_GEOM = (1 << 12), 498 /* disallow setting unmask bit */ 499 IDE_DFLAG_NO_UNMASK = (1 << 13), 500 /* disallow enabling 32-bit I/O */ 501 IDE_DFLAG_NO_IO_32BIT = (1 << 14), 502 /* for removable only: door lock/unlock works */ 503 IDE_DFLAG_DOORLOCKING = (1 << 15), 504 /* disallow DMA */ 505 IDE_DFLAG_NODMA = (1 << 16), 506 /* powermanagement told us not to do anything, so sleep nicely */ 507 IDE_DFLAG_BLOCKED = (1 << 17), 508 /* sleeping & sleep field valid */ 509 IDE_DFLAG_SLEEPING = (1 << 18), 510 IDE_DFLAG_POST_RESET = (1 << 19), 511 IDE_DFLAG_UDMA33_WARNED = (1 << 20), 512 IDE_DFLAG_LBA48 = (1 << 21), 513 /* status of write cache */ 514 IDE_DFLAG_WCACHE = (1 << 22), 515 /* used for ignoring ATA_DF */ 516 IDE_DFLAG_NOWERR = (1 << 23), 517 /* retrying in PIO */ 518 IDE_DFLAG_DMA_PIO_RETRY = (1 << 24), 519 IDE_DFLAG_LBA = (1 << 25), 520 /* don't unload heads */ 521 IDE_DFLAG_NO_UNLOAD = (1 << 26), 522 /* heads unloaded, please don't reset port */ 523 IDE_DFLAG_PARKED = (1 << 27), 524 IDE_DFLAG_MEDIA_CHANGED = (1 << 28), 525 /* write protect */ 526 IDE_DFLAG_WP = (1 << 29), 527 IDE_DFLAG_FORMAT_IN_PROGRESS = (1 << 30), 528 IDE_DFLAG_NIEN_QUIRK = (1 << 31), 529 }; 530 531 struct ide_drive_s { 532 char name[4]; /* drive name, such as "hda" */ 533 char driver_req[10]; /* requests specific driver */ 534 535 struct request_queue *queue; /* request queue */ 536 537 struct request *rq; /* current request */ 538 void *driver_data; /* extra driver data */ 539 u16 *id; /* identification info */ 540 #ifdef CONFIG_IDE_PROC_FS 541 struct proc_dir_entry *proc; /* /proc/ide/ directory entry */ 542 const struct ide_proc_devset *settings; /* /proc/ide/ drive settings */ 543 #endif 544 struct hwif_s *hwif; /* actually (ide_hwif_t *) */ 545 546 const struct ide_disk_ops *disk_ops; 547 548 unsigned long dev_flags; 549 550 unsigned long sleep; /* sleep until this time */ 551 unsigned long timeout; /* max time to wait for irq */ 552 553 u8 special_flags; /* special action flags */ 554 555 u8 select; /* basic drive/head select reg value */ 556 u8 retry_pio; /* retrying dma capable host in pio */ 557 u8 waiting_for_dma; /* dma currently in progress */ 558 u8 dma; /* atapi dma flag */ 559 560 u8 init_speed; /* transfer rate set at boot */ 561 u8 current_speed; /* current transfer rate set */ 562 u8 desired_speed; /* desired transfer rate set */ 563 u8 pio_mode; /* for ->set_pio_mode _only_ */ 564 u8 dma_mode; /* for ->set_dma_mode _only_ */ 565 u8 dn; /* now wide spread use */ 566 u8 acoustic; /* acoustic management */ 567 u8 media; /* disk, cdrom, tape, floppy, ... */ 568 u8 ready_stat; /* min status value for drive ready */ 569 u8 mult_count; /* current multiple sector setting */ 570 u8 mult_req; /* requested multiple sector setting */ 571 u8 io_32bit; /* 0=16-bit, 1=32-bit, 2/3=32bit+sync */ 572 u8 bad_wstat; /* used for ignoring ATA_DF */ 573 u8 head; /* "real" number of heads */ 574 u8 sect; /* "real" sectors per track */ 575 u8 bios_head; /* BIOS/fdisk/LILO number of heads */ 576 u8 bios_sect; /* BIOS/fdisk/LILO sectors per track */ 577 578 /* delay this long before sending packet command */ 579 u8 pc_delay; 580 581 unsigned int bios_cyl; /* BIOS/fdisk/LILO number of cyls */ 582 unsigned int cyl; /* "real" number of cyls */ 583 void *drive_data; /* used by set_pio_mode/dev_select() */ 584 unsigned int failures; /* current failure count */ 585 unsigned int max_failures; /* maximum allowed failure count */ 586 u64 probed_capacity;/* initial/native media capacity */ 587 u64 capacity64; /* total number of sectors */ 588 589 int lun; /* logical unit */ 590 int crc_count; /* crc counter to reduce drive speed */ 591 592 unsigned long debug_mask; /* debugging levels switch */ 593 594 #ifdef CONFIG_BLK_DEV_IDEACPI 595 struct ide_acpi_drive_link *acpidata; 596 #endif 597 struct list_head list; 598 struct device gendev; 599 struct completion gendev_rel_comp; /* to deal with device release() */ 600 601 /* current packet command */ 602 struct ide_atapi_pc *pc; 603 604 /* last failed packet command */ 605 struct ide_atapi_pc *failed_pc; 606 607 /* callback for packet commands */ 608 int (*pc_callback)(struct ide_drive_s *, int); 609 610 ide_startstop_t (*irq_handler)(struct ide_drive_s *); 611 612 unsigned long atapi_flags; 613 614 struct ide_atapi_pc request_sense_pc; 615 616 /* current sense rq and buffer */ 617 bool sense_rq_armed; 618 struct request *sense_rq; 619 struct request_sense sense_data; 620 }; 621 622 typedef struct ide_drive_s ide_drive_t; 623 624 #define to_ide_device(dev) container_of(dev, ide_drive_t, gendev) 625 626 #define to_ide_drv(obj, cont_type) \ 627 container_of(obj, struct cont_type, dev) 628 629 #define ide_drv_g(disk, cont_type) \ 630 container_of((disk)->private_data, struct cont_type, driver) 631 632 struct ide_port_info; 633 634 struct ide_tp_ops { 635 void (*exec_command)(struct hwif_s *, u8); 636 u8 (*read_status)(struct hwif_s *); 637 u8 (*read_altstatus)(struct hwif_s *); 638 void (*write_devctl)(struct hwif_s *, u8); 639 640 void (*dev_select)(ide_drive_t *); 641 void (*tf_load)(ide_drive_t *, struct ide_taskfile *, u8); 642 void (*tf_read)(ide_drive_t *, struct ide_taskfile *, u8); 643 644 void (*input_data)(ide_drive_t *, struct ide_cmd *, 645 void *, unsigned int); 646 void (*output_data)(ide_drive_t *, struct ide_cmd *, 647 void *, unsigned int); 648 }; 649 650 extern const struct ide_tp_ops default_tp_ops; 651 652 /** 653 * struct ide_port_ops - IDE port operations 654 * 655 * @init_dev: host specific initialization of a device 656 * @set_pio_mode: routine to program host for PIO mode 657 * @set_dma_mode: routine to program host for DMA mode 658 * @reset_poll: chipset polling based on hba specifics 659 * @pre_reset: chipset specific changes to default for device-hba resets 660 * @resetproc: routine to reset controller after a disk reset 661 * @maskproc: special host masking for drive selection 662 * @quirkproc: check host's drive quirk list 663 * @clear_irq: clear IRQ 664 * 665 * @mdma_filter: filter MDMA modes 666 * @udma_filter: filter UDMA modes 667 * 668 * @cable_detect: detect cable type 669 */ 670 struct ide_port_ops { 671 void (*init_dev)(ide_drive_t *); 672 void (*set_pio_mode)(struct hwif_s *, ide_drive_t *); 673 void (*set_dma_mode)(struct hwif_s *, ide_drive_t *); 674 blk_status_t (*reset_poll)(ide_drive_t *); 675 void (*pre_reset)(ide_drive_t *); 676 void (*resetproc)(ide_drive_t *); 677 void (*maskproc)(ide_drive_t *, int); 678 void (*quirkproc)(ide_drive_t *); 679 void (*clear_irq)(ide_drive_t *); 680 int (*test_irq)(struct hwif_s *); 681 682 u8 (*mdma_filter)(ide_drive_t *); 683 u8 (*udma_filter)(ide_drive_t *); 684 685 u8 (*cable_detect)(struct hwif_s *); 686 }; 687 688 struct ide_dma_ops { 689 void (*dma_host_set)(struct ide_drive_s *, int); 690 int (*dma_setup)(struct ide_drive_s *, struct ide_cmd *); 691 void (*dma_start)(struct ide_drive_s *); 692 int (*dma_end)(struct ide_drive_s *); 693 int (*dma_test_irq)(struct ide_drive_s *); 694 void (*dma_lost_irq)(struct ide_drive_s *); 695 /* below ones are optional */ 696 int (*dma_check)(struct ide_drive_s *, struct ide_cmd *); 697 int (*dma_timer_expiry)(struct ide_drive_s *); 698 void (*dma_clear)(struct ide_drive_s *); 699 /* 700 * The following method is optional and only required to be 701 * implemented for the SFF-8038i compatible controllers. 702 */ 703 u8 (*dma_sff_read_status)(struct hwif_s *); 704 }; 705 706 enum { 707 IDE_PFLAG_PROBING = (1 << 0), 708 }; 709 710 struct ide_host; 711 712 typedef struct hwif_s { 713 struct hwif_s *mate; /* other hwif from same PCI chip */ 714 struct proc_dir_entry *proc; /* /proc/ide/ directory entry */ 715 716 struct ide_host *host; 717 718 char name[6]; /* name of interface, eg. "ide0" */ 719 720 struct ide_io_ports io_ports; 721 722 unsigned long sata_scr[SATA_NR_PORTS]; 723 724 ide_drive_t *devices[MAX_DRIVES + 1]; 725 726 unsigned long port_flags; 727 728 u8 major; /* our major number */ 729 u8 index; /* 0 for ide0; 1 for ide1; ... */ 730 u8 channel; /* for dual-port chips: 0=primary, 1=secondary */ 731 732 u32 host_flags; 733 734 u8 pio_mask; 735 736 u8 ultra_mask; 737 u8 mwdma_mask; 738 u8 swdma_mask; 739 740 u8 cbl; /* cable type */ 741 742 hwif_chipset_t chipset; /* sub-module for tuning.. */ 743 744 struct device *dev; 745 746 void (*rw_disk)(ide_drive_t *, struct request *); 747 748 const struct ide_tp_ops *tp_ops; 749 const struct ide_port_ops *port_ops; 750 const struct ide_dma_ops *dma_ops; 751 752 /* dma physical region descriptor table (cpu view) */ 753 unsigned int *dmatable_cpu; 754 /* dma physical region descriptor table (dma view) */ 755 dma_addr_t dmatable_dma; 756 757 /* maximum number of PRD table entries */ 758 int prd_max_nents; 759 /* PRD entry size in bytes */ 760 int prd_ent_size; 761 762 /* Scatter-gather list used to build the above */ 763 struct scatterlist *sg_table; 764 int sg_max_nents; /* Maximum number of entries in it */ 765 766 struct ide_cmd cmd; /* current command */ 767 768 int rqsize; /* max sectors per request */ 769 int irq; /* our irq number */ 770 771 unsigned long dma_base; /* base addr for dma ports */ 772 773 unsigned long config_data; /* for use by chipset-specific code */ 774 unsigned long select_data; /* for use by chipset-specific code */ 775 776 unsigned long extra_base; /* extra addr for dma ports */ 777 unsigned extra_ports; /* number of extra dma ports */ 778 779 unsigned present : 1; /* this interface exists */ 780 unsigned busy : 1; /* serializes devices on a port */ 781 782 struct device gendev; 783 struct device *portdev; 784 785 struct completion gendev_rel_comp; /* To deal with device release() */ 786 787 void *hwif_data; /* extra hwif data */ 788 789 #ifdef CONFIG_BLK_DEV_IDEACPI 790 struct ide_acpi_hwif_link *acpidata; 791 #endif 792 793 /* IRQ handler, if active */ 794 ide_startstop_t (*handler)(ide_drive_t *); 795 796 /* BOOL: polling active & poll_timeout field valid */ 797 unsigned int polling : 1; 798 799 /* current drive */ 800 ide_drive_t *cur_dev; 801 802 /* current request */ 803 struct request *rq; 804 805 /* failsafe timer */ 806 struct timer_list timer; 807 /* timeout value during long polls */ 808 unsigned long poll_timeout; 809 /* queried upon timeouts */ 810 int (*expiry)(ide_drive_t *); 811 812 int req_gen; 813 int req_gen_timer; 814 815 spinlock_t lock; 816 } ____cacheline_internodealigned_in_smp ide_hwif_t; 817 818 #define MAX_HOST_PORTS 4 819 820 struct ide_host { 821 ide_hwif_t *ports[MAX_HOST_PORTS + 1]; 822 unsigned int n_ports; 823 struct device *dev[2]; 824 825 int (*init_chipset)(struct pci_dev *); 826 827 void (*get_lock)(irq_handler_t, void *); 828 void (*release_lock)(void); 829 830 irq_handler_t irq_handler; 831 832 unsigned long host_flags; 833 834 int irq_flags; 835 836 void *host_priv; 837 ide_hwif_t *cur_port; /* for hosts requiring serialization */ 838 839 /* used for hosts requiring serialization */ 840 volatile unsigned long host_busy; 841 }; 842 843 #define IDE_HOST_BUSY 0 844 845 /* 846 * internal ide interrupt handler type 847 */ 848 typedef ide_startstop_t (ide_handler_t)(ide_drive_t *); 849 typedef int (ide_expiry_t)(ide_drive_t *); 850 851 /* used by ide-cd, ide-floppy, etc. */ 852 typedef void (xfer_func_t)(ide_drive_t *, struct ide_cmd *, void *, unsigned); 853 854 extern struct mutex ide_setting_mtx; 855 856 /* 857 * configurable drive settings 858 */ 859 860 #define DS_SYNC (1 << 0) 861 862 struct ide_devset { 863 int (*get)(ide_drive_t *); 864 int (*set)(ide_drive_t *, int); 865 unsigned int flags; 866 }; 867 868 #define __DEVSET(_flags, _get, _set) { \ 869 .flags = _flags, \ 870 .get = _get, \ 871 .set = _set, \ 872 } 873 874 #define ide_devset_get(name, field) \ 875 static int get_##name(ide_drive_t *drive) \ 876 { \ 877 return drive->field; \ 878 } 879 880 #define ide_devset_set(name, field) \ 881 static int set_##name(ide_drive_t *drive, int arg) \ 882 { \ 883 drive->field = arg; \ 884 return 0; \ 885 } 886 887 #define ide_devset_get_flag(name, flag) \ 888 static int get_##name(ide_drive_t *drive) \ 889 { \ 890 return !!(drive->dev_flags & flag); \ 891 } 892 893 #define ide_devset_set_flag(name, flag) \ 894 static int set_##name(ide_drive_t *drive, int arg) \ 895 { \ 896 if (arg) \ 897 drive->dev_flags |= flag; \ 898 else \ 899 drive->dev_flags &= ~flag; \ 900 return 0; \ 901 } 902 903 #define __IDE_DEVSET(_name, _flags, _get, _set) \ 904 const struct ide_devset ide_devset_##_name = \ 905 __DEVSET(_flags, _get, _set) 906 907 #define IDE_DEVSET(_name, _flags, _get, _set) \ 908 static __IDE_DEVSET(_name, _flags, _get, _set) 909 910 #define ide_devset_rw(_name, _func) \ 911 IDE_DEVSET(_name, 0, get_##_func, set_##_func) 912 913 #define ide_devset_w(_name, _func) \ 914 IDE_DEVSET(_name, 0, NULL, set_##_func) 915 916 #define ide_ext_devset_rw(_name, _func) \ 917 __IDE_DEVSET(_name, 0, get_##_func, set_##_func) 918 919 #define ide_ext_devset_rw_sync(_name, _func) \ 920 __IDE_DEVSET(_name, DS_SYNC, get_##_func, set_##_func) 921 922 #define ide_decl_devset(_name) \ 923 extern const struct ide_devset ide_devset_##_name 924 925 ide_decl_devset(io_32bit); 926 ide_decl_devset(keepsettings); 927 ide_decl_devset(pio_mode); 928 ide_decl_devset(unmaskirq); 929 ide_decl_devset(using_dma); 930 931 #ifdef CONFIG_IDE_PROC_FS 932 /* 933 * /proc/ide interface 934 */ 935 936 #define ide_devset_rw_field(_name, _field) \ 937 ide_devset_get(_name, _field); \ 938 ide_devset_set(_name, _field); \ 939 IDE_DEVSET(_name, DS_SYNC, get_##_name, set_##_name) 940 941 #define ide_devset_rw_flag(_name, _field) \ 942 ide_devset_get_flag(_name, _field); \ 943 ide_devset_set_flag(_name, _field); \ 944 IDE_DEVSET(_name, DS_SYNC, get_##_name, set_##_name) 945 946 struct ide_proc_devset { 947 const char *name; 948 const struct ide_devset *setting; 949 int min, max; 950 int (*mulf)(ide_drive_t *); 951 int (*divf)(ide_drive_t *); 952 }; 953 954 #define __IDE_PROC_DEVSET(_name, _min, _max, _mulf, _divf) { \ 955 .name = __stringify(_name), \ 956 .setting = &ide_devset_##_name, \ 957 .min = _min, \ 958 .max = _max, \ 959 .mulf = _mulf, \ 960 .divf = _divf, \ 961 } 962 963 #define IDE_PROC_DEVSET(_name, _min, _max) \ 964 __IDE_PROC_DEVSET(_name, _min, _max, NULL, NULL) 965 966 typedef struct { 967 const char *name; 968 umode_t mode; 969 const struct file_operations *proc_fops; 970 } ide_proc_entry_t; 971 972 void proc_ide_create(void); 973 void proc_ide_destroy(void); 974 void ide_proc_register_port(ide_hwif_t *); 975 void ide_proc_port_register_devices(ide_hwif_t *); 976 void ide_proc_unregister_device(ide_drive_t *); 977 void ide_proc_unregister_port(ide_hwif_t *); 978 void ide_proc_register_driver(ide_drive_t *, struct ide_driver *); 979 void ide_proc_unregister_driver(ide_drive_t *, struct ide_driver *); 980 981 extern const struct file_operations ide_capacity_proc_fops; 982 extern const struct file_operations ide_geometry_proc_fops; 983 #else 984 static inline void proc_ide_create(void) { ; } 985 static inline void proc_ide_destroy(void) { ; } 986 static inline void ide_proc_register_port(ide_hwif_t *hwif) { ; } 987 static inline void ide_proc_port_register_devices(ide_hwif_t *hwif) { ; } 988 static inline void ide_proc_unregister_device(ide_drive_t *drive) { ; } 989 static inline void ide_proc_unregister_port(ide_hwif_t *hwif) { ; } 990 static inline void ide_proc_register_driver(ide_drive_t *drive, 991 struct ide_driver *driver) { ; } 992 static inline void ide_proc_unregister_driver(ide_drive_t *drive, 993 struct ide_driver *driver) { ; } 994 #endif 995 996 enum { 997 /* enter/exit functions */ 998 IDE_DBG_FUNC = (1 << 0), 999 /* sense key/asc handling */ 1000 IDE_DBG_SENSE = (1 << 1), 1001 /* packet commands handling */ 1002 IDE_DBG_PC = (1 << 2), 1003 /* request handling */ 1004 IDE_DBG_RQ = (1 << 3), 1005 /* driver probing/setup */ 1006 IDE_DBG_PROBE = (1 << 4), 1007 }; 1008 1009 /* DRV_NAME has to be defined in the driver before using the macro below */ 1010 #define __ide_debug_log(lvl, fmt, args...) \ 1011 { \ 1012 if (unlikely(drive->debug_mask & lvl)) \ 1013 printk(KERN_INFO DRV_NAME ": %s: " fmt "\n", \ 1014 __func__, ## args); \ 1015 } 1016 1017 /* 1018 * Power Management state machine (rq->pm->pm_step). 1019 * 1020 * For each step, the core calls ide_start_power_step() first. 1021 * This can return: 1022 * - ide_stopped : In this case, the core calls us back again unless 1023 * step have been set to ide_power_state_completed. 1024 * - ide_started : In this case, the channel is left busy until an 1025 * async event (interrupt) occurs. 1026 * Typically, ide_start_power_step() will issue a taskfile request with 1027 * do_rw_taskfile(). 1028 * 1029 * Upon reception of the interrupt, the core will call ide_complete_power_step() 1030 * with the error code if any. This routine should update the step value 1031 * and return. It should not start a new request. The core will call 1032 * ide_start_power_step() for the new step value, unless step have been 1033 * set to IDE_PM_COMPLETED. 1034 */ 1035 enum { 1036 IDE_PM_START_SUSPEND, 1037 IDE_PM_FLUSH_CACHE = IDE_PM_START_SUSPEND, 1038 IDE_PM_STANDBY, 1039 1040 IDE_PM_START_RESUME, 1041 IDE_PM_RESTORE_PIO = IDE_PM_START_RESUME, 1042 IDE_PM_IDLE, 1043 IDE_PM_RESTORE_DMA, 1044 1045 IDE_PM_COMPLETED, 1046 }; 1047 1048 int generic_ide_suspend(struct device *, pm_message_t); 1049 int generic_ide_resume(struct device *); 1050 1051 void ide_complete_power_step(ide_drive_t *, struct request *); 1052 ide_startstop_t ide_start_power_step(ide_drive_t *, struct request *); 1053 void ide_complete_pm_rq(ide_drive_t *, struct request *); 1054 void ide_check_pm_state(ide_drive_t *, struct request *); 1055 1056 /* 1057 * Subdrivers support. 1058 * 1059 * The gendriver.owner field should be set to the module owner of this driver. 1060 * The gendriver.name field should be set to the name of this driver 1061 */ 1062 struct ide_driver { 1063 const char *version; 1064 ide_startstop_t (*do_request)(ide_drive_t *, struct request *, sector_t); 1065 struct device_driver gen_driver; 1066 int (*probe)(ide_drive_t *); 1067 void (*remove)(ide_drive_t *); 1068 void (*resume)(ide_drive_t *); 1069 void (*shutdown)(ide_drive_t *); 1070 #ifdef CONFIG_IDE_PROC_FS 1071 ide_proc_entry_t * (*proc_entries)(ide_drive_t *); 1072 const struct ide_proc_devset * (*proc_devsets)(ide_drive_t *); 1073 #endif 1074 }; 1075 1076 #define to_ide_driver(drv) container_of(drv, struct ide_driver, gen_driver) 1077 1078 int ide_device_get(ide_drive_t *); 1079 void ide_device_put(ide_drive_t *); 1080 1081 struct ide_ioctl_devset { 1082 unsigned int get_ioctl; 1083 unsigned int set_ioctl; 1084 const struct ide_devset *setting; 1085 }; 1086 1087 int ide_setting_ioctl(ide_drive_t *, struct block_device *, unsigned int, 1088 unsigned long, const struct ide_ioctl_devset *); 1089 1090 int generic_ide_ioctl(ide_drive_t *, struct block_device *, unsigned, unsigned long); 1091 1092 extern int ide_vlb_clk; 1093 extern int ide_pci_clk; 1094 1095 int ide_end_rq(ide_drive_t *, struct request *, blk_status_t, unsigned int); 1096 void ide_kill_rq(ide_drive_t *, struct request *); 1097 1098 void __ide_set_handler(ide_drive_t *, ide_handler_t *, unsigned int); 1099 void ide_set_handler(ide_drive_t *, ide_handler_t *, unsigned int); 1100 1101 void ide_execute_command(ide_drive_t *, struct ide_cmd *, ide_handler_t *, 1102 unsigned int); 1103 1104 void ide_pad_transfer(ide_drive_t *, int, int); 1105 1106 ide_startstop_t ide_error(ide_drive_t *, const char *, u8); 1107 1108 void ide_fix_driveid(u16 *); 1109 1110 extern void ide_fixstring(u8 *, const int, const int); 1111 1112 int ide_busy_sleep(ide_drive_t *, unsigned long, int); 1113 1114 int __ide_wait_stat(ide_drive_t *, u8, u8, unsigned long, u8 *); 1115 int ide_wait_stat(ide_startstop_t *, ide_drive_t *, u8, u8, unsigned long); 1116 1117 ide_startstop_t ide_do_park_unpark(ide_drive_t *, struct request *); 1118 ide_startstop_t ide_do_devset(ide_drive_t *, struct request *); 1119 1120 extern ide_startstop_t ide_do_reset (ide_drive_t *); 1121 1122 extern int ide_devset_execute(ide_drive_t *drive, 1123 const struct ide_devset *setting, int arg); 1124 1125 void ide_complete_cmd(ide_drive_t *, struct ide_cmd *, u8, u8); 1126 int ide_complete_rq(ide_drive_t *, blk_status_t, unsigned int); 1127 1128 void ide_tf_readback(ide_drive_t *drive, struct ide_cmd *cmd); 1129 void ide_tf_dump(const char *, struct ide_cmd *); 1130 1131 void ide_exec_command(ide_hwif_t *, u8); 1132 u8 ide_read_status(ide_hwif_t *); 1133 u8 ide_read_altstatus(ide_hwif_t *); 1134 void ide_write_devctl(ide_hwif_t *, u8); 1135 1136 void ide_dev_select(ide_drive_t *); 1137 void ide_tf_load(ide_drive_t *, struct ide_taskfile *, u8); 1138 void ide_tf_read(ide_drive_t *, struct ide_taskfile *, u8); 1139 1140 void ide_input_data(ide_drive_t *, struct ide_cmd *, void *, unsigned int); 1141 void ide_output_data(ide_drive_t *, struct ide_cmd *, void *, unsigned int); 1142 1143 void SELECT_MASK(ide_drive_t *, int); 1144 1145 u8 ide_read_error(ide_drive_t *); 1146 void ide_read_bcount_and_ireason(ide_drive_t *, u16 *, u8 *); 1147 1148 int ide_check_ireason(ide_drive_t *, struct request *, int, int, int); 1149 1150 int ide_check_atapi_device(ide_drive_t *, const char *); 1151 1152 void ide_init_pc(struct ide_atapi_pc *); 1153 1154 /* Disk head parking */ 1155 extern wait_queue_head_t ide_park_wq; 1156 ssize_t ide_park_show(struct device *dev, struct device_attribute *attr, 1157 char *buf); 1158 ssize_t ide_park_store(struct device *dev, struct device_attribute *attr, 1159 const char *buf, size_t len); 1160 1161 /* 1162 * Special requests for ide-tape block device strategy routine. 1163 * 1164 * In order to service a character device command, we add special requests to 1165 * the tail of our block device request queue and wait for their completion. 1166 */ 1167 enum { 1168 REQ_IDETAPE_PC1 = (1 << 0), /* packet command (first stage) */ 1169 REQ_IDETAPE_PC2 = (1 << 1), /* packet command (second stage) */ 1170 REQ_IDETAPE_READ = (1 << 2), 1171 REQ_IDETAPE_WRITE = (1 << 3), 1172 }; 1173 1174 int ide_queue_pc_tail(ide_drive_t *, struct gendisk *, struct ide_atapi_pc *, 1175 void *, unsigned int); 1176 1177 int ide_do_test_unit_ready(ide_drive_t *, struct gendisk *); 1178 int ide_do_start_stop(ide_drive_t *, struct gendisk *, int); 1179 int ide_set_media_lock(ide_drive_t *, struct gendisk *, int); 1180 void ide_create_request_sense_cmd(ide_drive_t *, struct ide_atapi_pc *); 1181 void ide_retry_pc(ide_drive_t *drive); 1182 1183 void ide_prep_sense(ide_drive_t *drive, struct request *rq); 1184 int ide_queue_sense_rq(ide_drive_t *drive, void *special); 1185 1186 int ide_cd_expiry(ide_drive_t *); 1187 1188 int ide_cd_get_xferlen(struct request *); 1189 1190 ide_startstop_t ide_issue_pc(ide_drive_t *, struct ide_cmd *); 1191 1192 ide_startstop_t do_rw_taskfile(ide_drive_t *, struct ide_cmd *); 1193 1194 void ide_pio_bytes(ide_drive_t *, struct ide_cmd *, unsigned int, unsigned int); 1195 1196 void ide_finish_cmd(ide_drive_t *, struct ide_cmd *, u8); 1197 1198 int ide_raw_taskfile(ide_drive_t *, struct ide_cmd *, u8 *, u16); 1199 int ide_no_data_taskfile(ide_drive_t *, struct ide_cmd *); 1200 1201 int ide_taskfile_ioctl(ide_drive_t *, unsigned long); 1202 1203 int ide_dev_read_id(ide_drive_t *, u8, u16 *, int); 1204 1205 extern int ide_driveid_update(ide_drive_t *); 1206 extern int ide_config_drive_speed(ide_drive_t *, u8); 1207 extern u8 eighty_ninty_three (ide_drive_t *); 1208 extern int taskfile_lib_get_identify(ide_drive_t *drive, u8 *); 1209 1210 extern int ide_wait_not_busy(ide_hwif_t *hwif, unsigned long timeout); 1211 1212 extern void ide_stall_queue(ide_drive_t *drive, unsigned long timeout); 1213 1214 extern void ide_timer_expiry(unsigned long); 1215 extern irqreturn_t ide_intr(int irq, void *dev_id); 1216 extern void do_ide_request(struct request_queue *); 1217 extern void ide_requeue_and_plug(ide_drive_t *drive, struct request *rq); 1218 1219 void ide_init_disk(struct gendisk *, ide_drive_t *); 1220 1221 #ifdef CONFIG_IDEPCI_PCIBUS_ORDER 1222 extern int __ide_pci_register_driver(struct pci_driver *driver, struct module *owner, const char *mod_name); 1223 #define ide_pci_register_driver(d) __ide_pci_register_driver(d, THIS_MODULE, KBUILD_MODNAME) 1224 #else 1225 #define ide_pci_register_driver(d) pci_register_driver(d) 1226 #endif 1227 1228 static inline int ide_pci_is_in_compatibility_mode(struct pci_dev *dev) 1229 { 1230 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE && (dev->class & 5) != 5) 1231 return 1; 1232 return 0; 1233 } 1234 1235 void ide_pci_setup_ports(struct pci_dev *, const struct ide_port_info *, 1236 struct ide_hw *, struct ide_hw **); 1237 void ide_setup_pci_noise(struct pci_dev *, const struct ide_port_info *); 1238 1239 #ifdef CONFIG_BLK_DEV_IDEDMA_PCI 1240 int ide_pci_set_master(struct pci_dev *, const char *); 1241 unsigned long ide_pci_dma_base(ide_hwif_t *, const struct ide_port_info *); 1242 int ide_pci_check_simplex(ide_hwif_t *, const struct ide_port_info *); 1243 int ide_hwif_setup_dma(ide_hwif_t *, const struct ide_port_info *); 1244 #else 1245 static inline int ide_hwif_setup_dma(ide_hwif_t *hwif, 1246 const struct ide_port_info *d) 1247 { 1248 return -EINVAL; 1249 } 1250 #endif 1251 1252 struct ide_pci_enablebit { 1253 u8 reg; /* byte pci reg holding the enable-bit */ 1254 u8 mask; /* mask to isolate the enable-bit */ 1255 u8 val; /* value of masked reg when "enabled" */ 1256 }; 1257 1258 enum { 1259 /* Uses ISA control ports not PCI ones. */ 1260 IDE_HFLAG_ISA_PORTS = (1 << 0), 1261 /* single port device */ 1262 IDE_HFLAG_SINGLE = (1 << 1), 1263 /* don't use legacy PIO blacklist */ 1264 IDE_HFLAG_PIO_NO_BLACKLIST = (1 << 2), 1265 /* set for the second port of QD65xx */ 1266 IDE_HFLAG_QD_2ND_PORT = (1 << 3), 1267 /* use PIO8/9 for prefetch off/on */ 1268 IDE_HFLAG_ABUSE_PREFETCH = (1 << 4), 1269 /* use PIO6/7 for fast-devsel off/on */ 1270 IDE_HFLAG_ABUSE_FAST_DEVSEL = (1 << 5), 1271 /* use 100-102 and 200-202 PIO values to set DMA modes */ 1272 IDE_HFLAG_ABUSE_DMA_MODES = (1 << 6), 1273 /* 1274 * keep DMA setting when programming PIO mode, may be used only 1275 * for hosts which have separate PIO and DMA timings (ie. PMAC) 1276 */ 1277 IDE_HFLAG_SET_PIO_MODE_KEEP_DMA = (1 << 7), 1278 /* program host for the transfer mode after programming device */ 1279 IDE_HFLAG_POST_SET_MODE = (1 << 8), 1280 /* don't program host/device for the transfer mode ("smart" hosts) */ 1281 IDE_HFLAG_NO_SET_MODE = (1 << 9), 1282 /* trust BIOS for programming chipset/device for DMA */ 1283 IDE_HFLAG_TRUST_BIOS_FOR_DMA = (1 << 10), 1284 /* host is CS5510/CS5520 */ 1285 IDE_HFLAG_CS5520 = (1 << 11), 1286 /* ATAPI DMA is unsupported */ 1287 IDE_HFLAG_NO_ATAPI_DMA = (1 << 12), 1288 /* set if host is a "non-bootable" controller */ 1289 IDE_HFLAG_NON_BOOTABLE = (1 << 13), 1290 /* host doesn't support DMA */ 1291 IDE_HFLAG_NO_DMA = (1 << 14), 1292 /* check if host is PCI IDE device before allowing DMA */ 1293 IDE_HFLAG_NO_AUTODMA = (1 << 15), 1294 /* host uses MMIO */ 1295 IDE_HFLAG_MMIO = (1 << 16), 1296 /* no LBA48 */ 1297 IDE_HFLAG_NO_LBA48 = (1 << 17), 1298 /* no LBA48 DMA */ 1299 IDE_HFLAG_NO_LBA48_DMA = (1 << 18), 1300 /* data FIFO is cleared by an error */ 1301 IDE_HFLAG_ERROR_STOPS_FIFO = (1 << 19), 1302 /* serialize ports */ 1303 IDE_HFLAG_SERIALIZE = (1 << 20), 1304 /* host is DTC2278 */ 1305 IDE_HFLAG_DTC2278 = (1 << 21), 1306 /* 4 devices on a single set of I/O ports */ 1307 IDE_HFLAG_4DRIVES = (1 << 22), 1308 /* host is TRM290 */ 1309 IDE_HFLAG_TRM290 = (1 << 23), 1310 /* use 32-bit I/O ops */ 1311 IDE_HFLAG_IO_32BIT = (1 << 24), 1312 /* unmask IRQs */ 1313 IDE_HFLAG_UNMASK_IRQS = (1 << 25), 1314 IDE_HFLAG_BROKEN_ALTSTATUS = (1 << 26), 1315 /* serialize ports if DMA is possible (for sl82c105) */ 1316 IDE_HFLAG_SERIALIZE_DMA = (1 << 27), 1317 /* force host out of "simplex" mode */ 1318 IDE_HFLAG_CLEAR_SIMPLEX = (1 << 28), 1319 /* DSC overlap is unsupported */ 1320 IDE_HFLAG_NO_DSC = (1 << 29), 1321 /* never use 32-bit I/O ops */ 1322 IDE_HFLAG_NO_IO_32BIT = (1 << 30), 1323 /* never unmask IRQs */ 1324 IDE_HFLAG_NO_UNMASK_IRQS = (1 << 31), 1325 }; 1326 1327 #ifdef CONFIG_BLK_DEV_OFFBOARD 1328 # define IDE_HFLAG_OFF_BOARD 0 1329 #else 1330 # define IDE_HFLAG_OFF_BOARD IDE_HFLAG_NON_BOOTABLE 1331 #endif 1332 1333 struct ide_port_info { 1334 char *name; 1335 1336 int (*init_chipset)(struct pci_dev *); 1337 1338 void (*get_lock)(irq_handler_t, void *); 1339 void (*release_lock)(void); 1340 1341 void (*init_iops)(ide_hwif_t *); 1342 void (*init_hwif)(ide_hwif_t *); 1343 int (*init_dma)(ide_hwif_t *, 1344 const struct ide_port_info *); 1345 1346 const struct ide_tp_ops *tp_ops; 1347 const struct ide_port_ops *port_ops; 1348 const struct ide_dma_ops *dma_ops; 1349 1350 struct ide_pci_enablebit enablebits[2]; 1351 1352 hwif_chipset_t chipset; 1353 1354 u16 max_sectors; /* if < than the default one */ 1355 1356 u32 host_flags; 1357 1358 int irq_flags; 1359 1360 u8 pio_mask; 1361 u8 swdma_mask; 1362 u8 mwdma_mask; 1363 u8 udma_mask; 1364 }; 1365 1366 /* 1367 * State information carried for REQ_TYPE_ATA_PM_SUSPEND and REQ_TYPE_ATA_PM_RESUME 1368 * requests. 1369 */ 1370 struct ide_pm_state { 1371 /* PM state machine step value, currently driver specific */ 1372 int pm_step; 1373 /* requested PM state value (S1, S2, S3, S4, ...) */ 1374 u32 pm_state; 1375 void* data; /* for driver use */ 1376 }; 1377 1378 1379 int ide_pci_init_one(struct pci_dev *, const struct ide_port_info *, void *); 1380 int ide_pci_init_two(struct pci_dev *, struct pci_dev *, 1381 const struct ide_port_info *, void *); 1382 void ide_pci_remove(struct pci_dev *); 1383 1384 #ifdef CONFIG_PM 1385 int ide_pci_suspend(struct pci_dev *, pm_message_t); 1386 int ide_pci_resume(struct pci_dev *); 1387 #else 1388 #define ide_pci_suspend NULL 1389 #define ide_pci_resume NULL 1390 #endif 1391 1392 void ide_map_sg(ide_drive_t *, struct ide_cmd *); 1393 void ide_init_sg_cmd(struct ide_cmd *, unsigned int); 1394 1395 #define BAD_DMA_DRIVE 0 1396 #define GOOD_DMA_DRIVE 1 1397 1398 struct drive_list_entry { 1399 const char *id_model; 1400 const char *id_firmware; 1401 }; 1402 1403 int ide_in_drive_list(u16 *, const struct drive_list_entry *); 1404 1405 #ifdef CONFIG_BLK_DEV_IDEDMA 1406 int ide_dma_good_drive(ide_drive_t *); 1407 int __ide_dma_bad_drive(ide_drive_t *); 1408 1409 u8 ide_find_dma_mode(ide_drive_t *, u8); 1410 1411 static inline u8 ide_max_dma_mode(ide_drive_t *drive) 1412 { 1413 return ide_find_dma_mode(drive, XFER_UDMA_6); 1414 } 1415 1416 void ide_dma_off_quietly(ide_drive_t *); 1417 void ide_dma_off(ide_drive_t *); 1418 void ide_dma_on(ide_drive_t *); 1419 int ide_set_dma(ide_drive_t *); 1420 void ide_check_dma_crc(ide_drive_t *); 1421 ide_startstop_t ide_dma_intr(ide_drive_t *); 1422 1423 int ide_allocate_dma_engine(ide_hwif_t *); 1424 void ide_release_dma_engine(ide_hwif_t *); 1425 1426 int ide_dma_prepare(ide_drive_t *, struct ide_cmd *); 1427 void ide_dma_unmap_sg(ide_drive_t *, struct ide_cmd *); 1428 1429 #ifdef CONFIG_BLK_DEV_IDEDMA_SFF 1430 int config_drive_for_dma(ide_drive_t *); 1431 int ide_build_dmatable(ide_drive_t *, struct ide_cmd *); 1432 void ide_dma_host_set(ide_drive_t *, int); 1433 int ide_dma_setup(ide_drive_t *, struct ide_cmd *); 1434 extern void ide_dma_start(ide_drive_t *); 1435 int ide_dma_end(ide_drive_t *); 1436 int ide_dma_test_irq(ide_drive_t *); 1437 int ide_dma_sff_timer_expiry(ide_drive_t *); 1438 u8 ide_dma_sff_read_status(ide_hwif_t *); 1439 extern const struct ide_dma_ops sff_dma_ops; 1440 #else 1441 static inline int config_drive_for_dma(ide_drive_t *drive) { return 0; } 1442 #endif /* CONFIG_BLK_DEV_IDEDMA_SFF */ 1443 1444 void ide_dma_lost_irq(ide_drive_t *); 1445 ide_startstop_t ide_dma_timeout_retry(ide_drive_t *, int); 1446 1447 #else 1448 static inline u8 ide_find_dma_mode(ide_drive_t *drive, u8 speed) { return 0; } 1449 static inline u8 ide_max_dma_mode(ide_drive_t *drive) { return 0; } 1450 static inline void ide_dma_off_quietly(ide_drive_t *drive) { ; } 1451 static inline void ide_dma_off(ide_drive_t *drive) { ; } 1452 static inline void ide_dma_on(ide_drive_t *drive) { ; } 1453 static inline void ide_dma_verbose(ide_drive_t *drive) { ; } 1454 static inline int ide_set_dma(ide_drive_t *drive) { return 1; } 1455 static inline void ide_check_dma_crc(ide_drive_t *drive) { ; } 1456 static inline ide_startstop_t ide_dma_intr(ide_drive_t *drive) { return ide_stopped; } 1457 static inline ide_startstop_t ide_dma_timeout_retry(ide_drive_t *drive, int error) { return ide_stopped; } 1458 static inline void ide_release_dma_engine(ide_hwif_t *hwif) { ; } 1459 static inline int ide_dma_prepare(ide_drive_t *drive, 1460 struct ide_cmd *cmd) { return 1; } 1461 static inline void ide_dma_unmap_sg(ide_drive_t *drive, 1462 struct ide_cmd *cmd) { ; } 1463 #endif /* CONFIG_BLK_DEV_IDEDMA */ 1464 1465 #ifdef CONFIG_BLK_DEV_IDEACPI 1466 int ide_acpi_init(void); 1467 bool ide_port_acpi(ide_hwif_t *hwif); 1468 extern int ide_acpi_exec_tfs(ide_drive_t *drive); 1469 extern void ide_acpi_get_timing(ide_hwif_t *hwif); 1470 extern void ide_acpi_push_timing(ide_hwif_t *hwif); 1471 void ide_acpi_init_port(ide_hwif_t *); 1472 void ide_acpi_port_init_devices(ide_hwif_t *); 1473 extern void ide_acpi_set_state(ide_hwif_t *hwif, int on); 1474 #else 1475 static inline int ide_acpi_init(void) { return 0; } 1476 static inline bool ide_port_acpi(ide_hwif_t *hwif) { return 0; } 1477 static inline int ide_acpi_exec_tfs(ide_drive_t *drive) { return 0; } 1478 static inline void ide_acpi_get_timing(ide_hwif_t *hwif) { ; } 1479 static inline void ide_acpi_push_timing(ide_hwif_t *hwif) { ; } 1480 static inline void ide_acpi_init_port(ide_hwif_t *hwif) { ; } 1481 static inline void ide_acpi_port_init_devices(ide_hwif_t *hwif) { ; } 1482 static inline void ide_acpi_set_state(ide_hwif_t *hwif, int on) {} 1483 #endif 1484 1485 void ide_register_region(struct gendisk *); 1486 void ide_unregister_region(struct gendisk *); 1487 1488 void ide_check_nien_quirk_list(ide_drive_t *); 1489 void ide_undecoded_slave(ide_drive_t *); 1490 1491 void ide_port_apply_params(ide_hwif_t *); 1492 int ide_sysfs_register_port(ide_hwif_t *); 1493 1494 struct ide_host *ide_host_alloc(const struct ide_port_info *, struct ide_hw **, 1495 unsigned int); 1496 void ide_host_free(struct ide_host *); 1497 int ide_host_register(struct ide_host *, const struct ide_port_info *, 1498 struct ide_hw **); 1499 int ide_host_add(const struct ide_port_info *, struct ide_hw **, unsigned int, 1500 struct ide_host **); 1501 void ide_host_remove(struct ide_host *); 1502 int ide_legacy_device_add(const struct ide_port_info *, unsigned long); 1503 void ide_port_unregister_devices(ide_hwif_t *); 1504 void ide_port_scan(ide_hwif_t *); 1505 1506 static inline void *ide_get_hwifdata (ide_hwif_t * hwif) 1507 { 1508 return hwif->hwif_data; 1509 } 1510 1511 static inline void ide_set_hwifdata (ide_hwif_t * hwif, void *data) 1512 { 1513 hwif->hwif_data = data; 1514 } 1515 1516 extern void ide_toggle_bounce(ide_drive_t *drive, int on); 1517 1518 u64 ide_get_lba_addr(struct ide_cmd *, int); 1519 u8 ide_dump_status(ide_drive_t *, const char *, u8); 1520 1521 struct ide_timing { 1522 u8 mode; 1523 u8 setup; /* t1 */ 1524 u16 act8b; /* t2 for 8-bit io */ 1525 u16 rec8b; /* t2i for 8-bit io */ 1526 u16 cyc8b; /* t0 for 8-bit io */ 1527 u16 active; /* t2 or tD */ 1528 u16 recover; /* t2i or tK */ 1529 u16 cycle; /* t0 */ 1530 u16 udma; /* t2CYCTYP/2 */ 1531 }; 1532 1533 enum { 1534 IDE_TIMING_SETUP = (1 << 0), 1535 IDE_TIMING_ACT8B = (1 << 1), 1536 IDE_TIMING_REC8B = (1 << 2), 1537 IDE_TIMING_CYC8B = (1 << 3), 1538 IDE_TIMING_8BIT = IDE_TIMING_ACT8B | IDE_TIMING_REC8B | 1539 IDE_TIMING_CYC8B, 1540 IDE_TIMING_ACTIVE = (1 << 4), 1541 IDE_TIMING_RECOVER = (1 << 5), 1542 IDE_TIMING_CYCLE = (1 << 6), 1543 IDE_TIMING_UDMA = (1 << 7), 1544 IDE_TIMING_ALL = IDE_TIMING_SETUP | IDE_TIMING_8BIT | 1545 IDE_TIMING_ACTIVE | IDE_TIMING_RECOVER | 1546 IDE_TIMING_CYCLE | IDE_TIMING_UDMA, 1547 }; 1548 1549 struct ide_timing *ide_timing_find_mode(u8); 1550 u16 ide_pio_cycle_time(ide_drive_t *, u8); 1551 void ide_timing_merge(struct ide_timing *, struct ide_timing *, 1552 struct ide_timing *, unsigned int); 1553 int ide_timing_compute(ide_drive_t *, u8, struct ide_timing *, int, int); 1554 1555 #ifdef CONFIG_IDE_XFER_MODE 1556 int ide_scan_pio_blacklist(char *); 1557 const char *ide_xfer_verbose(u8); 1558 int ide_pio_need_iordy(ide_drive_t *, const u8); 1559 int ide_set_pio_mode(ide_drive_t *, u8); 1560 int ide_set_dma_mode(ide_drive_t *, u8); 1561 void ide_set_pio(ide_drive_t *, u8); 1562 int ide_set_xfer_rate(ide_drive_t *, u8); 1563 #else 1564 static inline void ide_set_pio(ide_drive_t *drive, u8 pio) { ; } 1565 static inline int ide_set_xfer_rate(ide_drive_t *drive, u8 rate) { return -1; } 1566 #endif 1567 1568 static inline void ide_set_max_pio(ide_drive_t *drive) 1569 { 1570 ide_set_pio(drive, 255); 1571 } 1572 1573 char *ide_media_string(ide_drive_t *); 1574 1575 extern const struct attribute_group *ide_dev_groups[]; 1576 extern struct bus_type ide_bus_type; 1577 extern struct class *ide_port_class; 1578 1579 static inline void ide_dump_identify(u8 *id) 1580 { 1581 print_hex_dump(KERN_INFO, "", DUMP_PREFIX_NONE, 16, 2, id, 512, 0); 1582 } 1583 1584 static inline int hwif_to_node(ide_hwif_t *hwif) 1585 { 1586 return hwif->dev ? dev_to_node(hwif->dev) : -1; 1587 } 1588 1589 static inline ide_drive_t *ide_get_pair_dev(ide_drive_t *drive) 1590 { 1591 ide_drive_t *peer = drive->hwif->devices[(drive->dn ^ 1) & 1]; 1592 1593 return (peer->dev_flags & IDE_DFLAG_PRESENT) ? peer : NULL; 1594 } 1595 1596 static inline void *ide_get_drivedata(ide_drive_t *drive) 1597 { 1598 return drive->drive_data; 1599 } 1600 1601 static inline void ide_set_drivedata(ide_drive_t *drive, void *data) 1602 { 1603 drive->drive_data = data; 1604 } 1605 1606 #define ide_port_for_each_dev(i, dev, port) \ 1607 for ((i) = 0; ((dev) = (port)->devices[i]) || (i) < MAX_DRIVES; (i)++) 1608 1609 #define ide_port_for_each_present_dev(i, dev, port) \ 1610 for ((i) = 0; ((dev) = (port)->devices[i]) || (i) < MAX_DRIVES; (i)++) \ 1611 if ((dev)->dev_flags & IDE_DFLAG_PRESENT) 1612 1613 #define ide_host_for_each_port(i, port, host) \ 1614 for ((i) = 0; ((port) = (host)->ports[i]) || (i) < MAX_HOST_PORTS; (i)++) 1615 1616 1617 #endif /* _IDE_H */ 1618
Linux® is a registered trademark of Linus Torvalds in the United States and other countries.
TOMOYO® is a registered trademark of NTT DATA CORPORATION.