From 7a7cf0355140357aafc48cbcf79b3b0dc1e574cd Mon Sep 17 00:00:00 2001 From: Thomas White Date: Wed, 6 Jul 2011 15:53:31 +0200 Subject: Fix the choice of "guiding" reflections for PR --- src/geometry.c | 2 +- src/geometry.h | 2 +- src/hrs-scaling.c | 183 ++++++++++++++++++++++++++++++++++---------------- src/hrs-scaling.h | 6 +- src/partialator.c | 167 ++++++++++++++++++++++----------------------- src/post-refinement.c | 36 +++++----- src/post-refinement.h | 3 +- 7 files changed, 225 insertions(+), 174 deletions(-) diff --git a/src/geometry.c b/src/geometry.c index df1e8ac4..b338da56 100644 --- a/src/geometry.c +++ b/src/geometry.c @@ -347,7 +347,7 @@ void predict_corresponding_reflections(struct image *image, const char *sym, /* Calculate partialities and apply them to the image's raw_reflections */ -void update_partialities(struct image *image, const char *sym) +void update_partialities(struct image *image) { Reflection *refl; RefListIterator *iter; diff --git a/src/geometry.h b/src/geometry.h index efda6c87..c47f4a27 100644 --- a/src/geometry.h +++ b/src/geometry.h @@ -25,7 +25,7 @@ extern void predict_corresponding_reflections(struct image *image, const char *sym, int *n_expected, int *n_found, int *n_notfound); -extern void update_partialities(struct image *image, const char *sym); +extern void update_partialities(struct image *image); #endif /* GEOMETRY_H */ diff --git a/src/hrs-scaling.c b/src/hrs-scaling.c index 2e2f8d99..ff73ee77 100644 --- a/src/hrs-scaling.c +++ b/src/hrs-scaling.c @@ -35,7 +35,7 @@ #define MAX_CYCLES (50) -char *find_common_reflections(struct image *images, int n) +static char *find_common_reflections(struct image *images, int n) { int i; char *cref; @@ -54,6 +54,8 @@ char *find_common_reflections(struct image *images, int n) signed int h, k, l; int j; + if ( !get_scalable(refl) ) continue; + get_indices(refl, &h, &k, &l); for ( j=0; jh, it->k, it->l, sym, &uh, &vh); + get_indices(refl, &h, &k, &l); - set_intensity(uh_arr, it->h, it->k, it->l, uh); - set_intensity(vh_arr, it->h, it->k, it->l, vh); + s_uhvh(images, n, h, k, l, &uh, &vh); + set_intensity(uh_arr, h, k, l, uh); + set_intensity(vh_arr, h, k, l, vh); } uha_arr = malloc(n*sizeof(double)); vha_arr = malloc(n*sizeof(double)); - for ( refidx=0; refidxh; - const signed int k = it->k; - const signed int l = it->l; + signed int h, k, l; + + get_indices(refl, &h, &k, &l); /* For this reflection, calculate all the possible * values of uha and vha */ @@ -290,7 +292,15 @@ static double iterate_scale(struct image *images, int n, if ( isnan(Ih) ) Ih = 0.0; } else { /* Look up by asymmetric indices */ - Ih = lookup_intensity(reference, h, k, l); + Reflection *r = find_refl(reference, h, k, l); + if ( r == NULL ) { + ERROR("%3i %3i %3i isn't in the " + "reference list, so why is it " + "marked as scalable?\n", h, k, l); + Ih = 0.0; + } else { + Ih = get_intensity(r); + } } rha = vha - image_a->osf * uha * Ih; @@ -310,7 +320,9 @@ static double iterate_scale(struct image *images, int n, if ( (reference != NULL) && (a!=b) ) continue; /* Zero if no common reflections */ - if ( cref[a+n*b] != 0 ) continue; + if ( (reference == NULL) && cref[a+n*b] != 0 ) { + continue; + } uhb = uha_arr[b]; vhb = vha_arr[b]; @@ -335,7 +347,6 @@ static double iterate_scale(struct image *images, int n, gsl_vector_set(v, a, vval+vc); } - } free(uh_arr); @@ -379,27 +390,28 @@ static double iterate_scale(struct image *images, int n, } -static RefList *lsq_intensities(struct image *images, int n, - ReflItemList *obs, const char *sym) +static void lsq_intensities(struct image *images, int n, RefList *full) { - RefList *full; - int i; - - full = reflist_new(); - for ( i=0; ih, it->k, it->l); - refl != NULL; - refl = next_found_refl(refl) ) + for ( f = find_refl(images[m].reflections, h, k, l); + f != NULL; f = next_found_refl(refl) ) { double p; - if ( !get_scalable(refl) ) continue; - p = get_partiality(refl); + if ( !get_scalable(f) ) continue; + p = get_partiality(f); - num += get_intensity(refl) * p * G; + num += get_intensity(f) * p * G; den += pow(p, 2.0) * pow(G, 2.0); redundancy++; @@ -426,19 +436,17 @@ static RefList *lsq_intensities(struct image *images, int n, } - if ( !isnan(num/den) ) { - new = add_refl(full, it->h, it->k, it->l); - set_int(new, num/den); - set_redundancy(new, redundancy); - } else { - ERROR("Couldn't calculate LSQ full intensity for" - "%3i %3i %3i\n", it->h, it->k, it->l); - /* Doom is probably impending */ + set_int(refl, num/den); + + //STATUS("%3i %3i %3i %i %i\n", h, k, l, + // redundancy, get_redundancy(refl)); + + if ( get_redundancy(refl) != redundancy ) { + ERROR("Didn't find all the expected parts for" + " %3i %3i %3i\n", h, k, l); } } - - return full; } @@ -459,23 +467,78 @@ static void normalise_osfs(struct image *images, int n) } +static RefList *guess_scaled_reflections(struct image *images, int n) +{ + RefList *scalable; + int i; + + scalable = reflist_new(); + for ( i=0; i 0.01) && (i < MAX_CYCLES) ); - full = lsq_intensities(images, n, obs, sym); + free(cref); + + show_scale_factors(images, n); + + lsq_intensities(images, n, full); return full; } diff --git a/src/hrs-scaling.h b/src/hrs-scaling.h index f65b7988..a18d3a0b 100644 --- a/src/hrs-scaling.h +++ b/src/hrs-scaling.h @@ -20,10 +20,8 @@ #include "image.h" -extern RefList *scale_intensities(struct image *images, int n, const char *sym, - ReflItemList *obs, char *cref, - double *reference); +extern RefList *scale_intensities(struct image *images, int n, + RefList *reference); -extern char *find_common_reflections(struct image *images, int n); #endif /* HRS_SCALING_H */ diff --git a/src/partialator.c b/src/partialator.c index 9e09b4c7..252d6cfd 100644 --- a/src/partialator.c +++ b/src/partialator.c @@ -66,8 +66,6 @@ static void show_help(const char *s) struct refine_args { - const char *sym; - ReflItemList *obs; RefList *full; struct image *image; FILE *graph; @@ -91,7 +89,7 @@ static void refine_image(void *task, int id) struct image *image = pargs->image; image->id = id; - pr_refine(image, pargs->full, pargs->sym); + pr_refine(image, pargs->full); } @@ -123,15 +121,13 @@ static void done_image(void *vqargs, void *task) static void refine_all(struct image *images, int n_total_patterns, - struct detector *det, const char *sym, - ReflItemList *obs, RefList *full, int nthreads, + struct detector *det, + RefList *full, int nthreads, FILE *graph, FILE *pgraph) { struct refine_args task_defaults; struct queue_args qargs; - task_defaults.sym = sym; - task_defaults.obs = obs; task_defaults.full = full; task_defaults.image = NULL; task_defaults.graph = graph; @@ -152,7 +148,7 @@ static void refine_all(struct image *images, int n_total_patterns, /* Decide which reflections can be scaled */ -static int select_scalable_reflections(RefList *list, ReflItemList *sc_l) +static int select_scalable_reflections(RefList *list, RefList *reference) { Reflection *refl; RefListIterator *iter; @@ -162,29 +158,24 @@ static int select_scalable_reflections(RefList *list, ReflItemList *sc_l) refl != NULL; refl = next_refl(refl, iter) ) { - int scalable = 1; + int sc = 1; double v; - if ( get_partiality(refl) < 0.1 ) scalable = 0; + if ( get_partiality(refl) < 0.1 ) sc = 0; v = fabs(get_intensity(refl)); - if ( v < 0.1 ) scalable = 0; - set_scalable(refl, scalable); - - if ( scalable ) { + if ( v < 0.1 ) sc = 0; + /* If we are scaling against a reference set, we additionally + * require that this reflection is in the reference list. */ + if ( reference != NULL ) { signed int h, k, l; - - nobs++; - - /* Add (asymmetric) indices to list */ get_indices(refl, &h, &k, &l); - - if ( !find_item(sc_l, h, k, l) ) { - add_item(sc_l, h, k, l); - } - + if ( find_refl(reference, h, k, l) == NULL ) sc = 0; } + set_scalable(refl, sc); + + if ( sc ) nobs++; } return nobs; @@ -192,7 +183,7 @@ static int select_scalable_reflections(RefList *list, ReflItemList *sc_l) static void select_reflections_for_refinement(struct image *images, int n, - ReflItemList *scalable) + RefList *full, int have_reference) { int i; @@ -200,6 +191,11 @@ static void select_reflections_for_refinement(struct image *images, int n, Reflection *refl; RefListIterator *iter; + int n_acc = 0; + int n_nomatch = 0; + int n_noscale = 0; + int n_fewmatch = 0; + int n_ref = 0; for ( refl = first_refl(images[i].reflections, &iter); refl != NULL; @@ -208,14 +204,47 @@ static void select_reflections_for_refinement(struct image *images, int n, signed int h, k, l; int sc; + n_ref++; + + /* We require that the reflection itself is scalable + * (i.e. sensible partiality and intensity) and that + * the "full" estimate of this reflection is made from + * at least two parts. */ get_indices(refl, &h, &k, &l); sc = get_scalable(refl); + if ( !sc ) { + + n_noscale++; + + } else { + + Reflection *f = find_refl(full, h, k, l); + + if ( f != NULL ) { + + int r = get_redundancy(f); + if ( (r >= 2) || have_reference ) { + set_refinable(refl, 1); + n_acc++; + } else { + n_fewmatch++; + } + + } else { + n_nomatch++; + } - if ( sc && find_item(scalable, h, k, l) ) { - set_refinable(refl, 1); } } + STATUS("Image %4i: %i guide reflections accepted " + "(%i not scalable, %i few matches, %i total)\n", + i, n_acc, n_noscale, n_fewmatch, n_ref); + + /* This would be a silly situation, since there must be a match + * if THIS pattern has a scalable part of the reflection! */ + assert(n_nomatch == 0); + } } @@ -230,7 +259,6 @@ int main(int argc, char *argv[]) FILE *fh; int nthreads = 1; struct detector *det; - ReflItemList *scalable; int i; int n_total_patterns; struct image *images; @@ -240,13 +268,12 @@ int main(int argc, char *argv[]) int n_found = 0; int n_expected = 0; int n_notfound = 0; - char *cref; int n_usable_patterns = 0; int nobs; char *reference_file = NULL; - double *reference = NULL; - RefList *reference_list = NULL; + RefList *reference = NULL; int n_dud; + int have_reference = 0; /* Long options */ const struct option longopts[] = { @@ -359,9 +386,9 @@ int main(int argc, char *argv[]) list = read_reflections(reference_file); free(reference_file); if ( list == NULL ) return 1; - reference_list = asymmetric_indices(list, sym); + reference = asymmetric_indices(list, sym); reflist_free(list); - reference = intensities_from_list(reference_list); + have_reference = 1; } @@ -382,7 +409,6 @@ int main(int argc, char *argv[]) /* Fill in what we know about the images so far */ rewind(fh); - scalable = new_items(); nobs = 0; for ( i=0; ireflections = as; predict_corresponding_reflections(cur, sym, &n_expected, - &n_found, &n_notfound); + &n_found, &n_notfound); - nobs += select_scalable_reflections(cur->reflections, scalable); + nobs += select_scalable_reflections(cur->reflections, + reference); progress_bar(i, n_total_patterns-1, "Loading pattern data"); n_usable_patterns++; @@ -438,51 +465,13 @@ int main(int argc, char *argv[]) fclose(fh); STATUS("Found %5.2f%% of the expected peaks (missed %i of %i).\n", 100.0 * (double)n_found / n_expected, n_notfound, n_expected); - STATUS("Mean measurements per scalable unique reflection: %5.2f\n", - (double)nobs / num_items(scalable)); - - cref = find_common_reflections(images, n_usable_patterns); /* Make initial estimates */ STATUS("Performing initial scaling.\n"); - full = scale_intensities(images, n_usable_patterns, sym, - scalable, cref, reference); - - select_reflections_for_refinement(images, n_usable_patterns, scalable); - - for ( i=0; ih, it->k, it->l); - if ( f == NULL ) { - ERROR("%3i %3i %3i was designated scalable, but no" - " full intensity was recorded.\n", - it->h, it->k, it->l); - } - } + full = scale_intensities(images, n_usable_patterns, reference); - for ( i=0; ireflections, - scalable); + reference); } - STATUS("Mean measurements per scalable unique " - "reflection: %5.2f\n", (double)nobs/num_items(scalable)); /* Re-estimate all the full intensities */ reflist_free(full); full = scale_intensities(images, n_usable_patterns, - sym, scalable, cref, reference); + reference); + + select_reflections_for_refinement(images, n_usable_patterns, + full, have_reference); fclose(fhg); fclose(fhp); @@ -557,15 +551,12 @@ int main(int argc, char *argv[]) reflist_free(images[i].reflections); } reflist_free(full); - delete_items(scalable); free(sym); free(outfile); free_detector_geometry(det); free(beam); - free(cref); if ( reference != NULL ) { - free(reference); - reflist_free(reference_list); + reflist_free(reference); } for ( i=0; iosf * get_intensity(match); /* Actual measurement of this reflection from this pattern? */ @@ -402,9 +402,9 @@ static double pr_iterate(struct image *image, const RefList *full, } //show_matrix_eqn(M, v, NUM_PARAMS); - //STATUS("%i reflections were scalable\n", nref); + //STATUS("%i reflections went into the equations.\n", nref); if ( nref == 0 ) { - ERROR("No reflections left to scale!\n"); + ERROR("No guide reflections to refine with!\n"); gsl_matrix_free(M); gsl_vector_free(v); return 0.0; @@ -434,8 +434,7 @@ static double pr_iterate(struct image *image, const RefList *full, } -static double mean_partial_dev(struct image *image, - const RefList *full, const char *sym) +static double guide_dev(struct image *image, const RefList *full) { double dev = 0.0; @@ -452,13 +451,11 @@ static double mean_partial_dev(struct image *image, Reflection *full_version; double I_full, I_partial; - if ( !get_scalable(refl) ) continue; + if ( !get_refinable(refl) ) continue; get_indices(refl, &h, &k, &l); assert((h!=0) || (k!=0) || (l!=0)); - if ( !get_scalable(refl) ) continue; - full_version = find_refl(full, h, k, l); if ( full_version == NULL ) continue; /* Some reflections may have recently become scalable, but @@ -481,14 +478,14 @@ static double mean_partial_dev(struct image *image, } -void pr_refine(struct image *image, const RefList *full, const char *sym) +void pr_refine(struct image *image, const RefList *full) { double max_shift, dev; int i; const int verbose = 1; if ( verbose ) { - dev = mean_partial_dev(image, full, sym); + dev = guide_dev(image, full); STATUS("\n"); /* Deal with progress bar */ STATUS("Before iteration: dev = %10.5e\n", dev); @@ -506,15 +503,14 @@ void pr_refine(struct image *image, const RefList *full, const char *sym) cell_get_reciprocal(image->indexed_cell, &asx, &asy, &asz, &bsx, &bsy, &bsz, &csx, &csy, &csz); - max_shift = pr_iterate(image, full, sym); + max_shift = pr_iterate(image, full); - update_partialities(image, sym); + update_partialities(image); if ( verbose ) { - dev = mean_partial_dev(image, full, sym); + dev = guide_dev(image, full); STATUS("PR Iteration %2i: max shift = %10.2f" - " dev = %10.5e\n", - i+1, max_shift, dev); + " dev = %10.5e\n", i+1, max_shift, dev); } i++; diff --git a/src/post-refinement.h b/src/post-refinement.h index 039a69c8..28d65877 100644 --- a/src/post-refinement.h +++ b/src/post-refinement.h @@ -41,8 +41,7 @@ enum { }; -extern void pr_refine(struct image *image, const RefList *full, - const char *sym); +extern void pr_refine(struct image *image, const RefList *full); /* Exported so it can be poked by tests/pr_gradient_check */ extern double gradient(struct image *image, int k, Reflection *refl, double r); -- cgit v1.2.3