Category Archives: Knowledge Ecology And Economy

World Health Day

‘Working together for health’ is this year’s slogan for World Health Day (today).
The World Health Organisation (WHO) quite rightly asks that we take time just for one day in the year to think about what ‘Health’ actually means. So today, 7 April 2006, is World Health Day.
This year’s strapline is ‘working together for health’. Reduction in child mortality, improvements in maternal health and combatting HIV / AIDS, malaria and other diseases are amongst the Millennium Development Goals ** which all Member States are signed up to meet by the year 2015.
[** The other five goals are eradication of extreme poverty and hunger, universal primary education, promotion of gender equality and the empowerment of women, environmental sustainability and a global parnership for development.]
A time to reflect?
But here we are in Britain, one of the half-dozen most wealthy countries in the world, and even we don’t get it right on all counts. There is plenty to worry about in the health of our nation; but it cannot be said, as of some other countries, that anyone ‘has’ to starve or die of cold, lack of clean water or because of any of the other horrendous experiences of people in other parts of the world. In the U.K. we have choices, and we have resources, which really do mean this never has to happen.
I say this neither (I hope) to make inappropriate comparisons – poverty in health or anything else in the U.K. is relative; poverty elsewhere is grimly absolute – nor to offer bland pronouncements about what we ‘ought’ to do to reduce such awful suffering in other areas of the globe.
What I seek to understand more clearly is how we can think in a more joined-up way.
Only connect…
We in Britain, like those in other first world countries, mostly know that how we treat our bodies and what we do to promote sustainability are critical both for ourselves and to what happens to people elsewhere, as well as people here. We know too that responsibility for this lies with us personally and as parents, as well as with ‘the government’, or ‘them’.
There’s a message here. It’s at base very simple. The fundamental question is, how do we deliver action?
If World Health Day does nothing else, perhaps it encourages us to reflect how, across the globe, we are all interconnected and interdependent. The ‘links’ are there, on the internet and, even more importantly, in our hearts and minds.

The Dismal Message Of Some Human Resource Advisers

Human resource specialists seem to spend a lot of time these days developing ways of ‘testing’ potential employees. Technology does have a part to play in assessing candidiates for jobs, not least because it comprises an attempt to move beyond stereoypical and unfair assumptions. But to work to greatest effect technologically-led assessment must be considered carefully, and with due acknowledgement of the difficulties of ‘proving’ it is meaningful. If educators made the same deterministic (and dubious) assumptions as some human resource managers, there would be far less call for educational services.
Like everything else in our brave new technological world, ‘human resourcing’ has been re-branded as a science.
In many ways, this is to be applauded. Anything which moves us on from the old-style way of ‘jobs for the boys’ has to be an improvement. But in at least one respect it’s worrying.
New-style appointment procedures
Many appointment procedures now begin with an application form which asks questions about how the propective employee has tackled a variety of challenges, followed by ‘assessment’ at a ‘centre’. And only after all that does a real human being perhaps deign to conduct a personal conversation or interview with candidates about the post in question.
So, once the standard information has been recorded, the initial application form in these cases often states that in one way or another that ‘past experience is the best indicator of future performance’. This is generally the prelude to a requirement that the applicant gives ‘brief’ accounts of how he or she dealt with a difficult situation, resolved a dilemma or took a fractious group of people to some sort of resolution of their problems.
Real scenarios, or imagined?
Wonderful. Presumably in every case the job the applicant is going for requires a vivid imagination? Because the only sure thing we can learn from such accounts is that people are good – or not – at writing (very) short stories. These may indeed be stories related to the skills and scenarios of the post in question, but they are hardly testable against hard evidence.
Degree certificates may not tell us much, but they do confirm that a job applicant’s claim to be a graduate (or whatever) is genuine. Such genuineness cannot be established for these ‘mini stories’. Of course many people do tell the truth when they give accounts of their past actions; but it’s a certainty that not all of them will be doing so.
Thus we are confronted by a situation in which those who stick firmly to the truth (as they understand it) are likely to be competing against others who are not so fussy about such matters… or who simply have convenient memories. In this context, what useful value can anyone put on the assumption that person’s future actions will follow from their reported history?
Assessing what?
Then there’s the second stage of the selection process. For some posts it is fair enough to ask the candidate to perform tasks like the one s/he aspires to in the job on offer. If you say you can type forty words a minute, then here’s your chance to demonstrate that. If you claim to be able to work with spreadsheets, please go ahead and show you can.
But some tasks, especially at more senior levels, require careful and balanced judgement. They are about bringing experience and human insight to bear on difficult situations. They require a wide grasp of the influencing factors and a steadfastness in terms of dealing with people.
I.Q. tests by another name?
In such tasks there is little reason to suppose, as increasingly is assumed, that non-contextualised ‘verbal reasoning’ tests and the like will take us very far. An untitled poorly written paragraph of general assertions such as often appears as part of these ‘tests’, giving no indication of who wrote it, or for whom, makes little sense to those who know that all real-world interaction takes place in the context of unarticulated as well as formal intentions. This real-life exerience makes it pretty problematic to answer stark multiple choice (i.e. non-discursive, computer-markable) questions about what such paragraphs supposed to ‘mean’.
Perhaps there’s an irony, given the observations above, in the likelihood that only those without much imagination may be comfortable responding to such mechanistic examination.
The validity of I.Q. (Intelligence Quotient) tests has been under challenge for a full half century now; and the British ’11 Plus’ examination – in some respects a precursor of these types of mechanistic tets – is rightly history, discredited and still a cause of great distress to many whom it so cruelly labelled unsuitable for more academic or rigourous secondary education.
The clash between the premises of education and those of candidiate selection
Checking out the general abilitiies of people going for particular jobs is a sensible idea. Myers Briggs tests, for instance, give a useful – though not infallable – indication of how a person might approach given situations, thus helping everyone (including the candidate) to assess whether s/he is ‘right for the job’.
This, however, is a long way from the claim that what’s happened in the past is a good indicator of a perons’s future behaviour and capability. Indeed, those of us who have worked in adult (and perhaps even children’s) education would be out of business entirely if everyone believed this to be so.
Of course past experience helps to mould future behaviour; but much more important, given a general level of aptitude and attitude, is the opportunity which presents, or does not present, to learn and develop.
How do they know?
It’s a question which I’ve asked before, but it still seems reasonable: How do the decision-makers actually know that their way of doing things selects the best applicants? And the answer is, unless they’ve followed up those who weren’t successful, and compared them (using fully valid criteria) with those who were, they cannot know for sure.
The human resource specialists and assessment centre gurus may be covering their backs and keeping some employers happy whilst they’re at it. And no doubt some do a very good job. But it still seems indefensible to claim that they really know what potential employees are ‘like’ on the basis of their new-style forms and some of the tests which have been devised for selection.
Those of us who have more belief in the capacity of people to grow and learn might fear the new ‘science’ of human resourcing sometimes gets dangerously close to the dismally deterministic ways of the discredited ‘educationalists’ of yore.
Frankly, were I a prospective employer, I’d expect more for my human resource investment than that.

Time Is Energy (And ‘Clocks Forward’ Daylight Uses Less)

Dawn (small).jpg‘Daylight saving’ is a strange notion. But ‘daylight energy saving’ is a very different consideration. How we arrange the hours of light and darkness across our working day has many impacts – which makes it all the more curious that so little high profile or current research has been focused on British Summer Time and rationales for why the clocks ‘go back’ in the Winter.
My recent piece on British Summer Time has drawn a lot of comment, both on and off this website.
Highway at night.jpgThere are people who seem simply not to mind whether / when it’s light or dark as they go about their daily business, but there are many others who have responded quite strongly in terms of their need for as much daylight as possible. It must be very helpful in some ways not to mind how dark it is, but it’s quite incomprehensible to others that there are folk who genuinely ‘don’t mind’. Perhaps it’s rather like being ‘colour blind’ – if you don’t perceive the difference between red and green you just accept (and may not even know about) it; or maybe some of us have physiologies which are more photo-sensitive than others.
Daylight saving is energy saving
The most important thing to come out of the discussions so far, however, is not that people may have different personal preferences, but that the terms of engagement in this debate are becoming clearer.
One striking aspect of so-called ‘daylight saving’ which is emerging, alongside the prime safety considerations, is its significance not only potentially for business efficiency, but also, even more crucially, for energy. It does begin to look very much as though more ‘summertime’ would keep energy consumption down.
Where’s the evidence?
A big surprise in all this is the paucity of serious publicly available evidence other than on safety (avoidance of accidents). It seems in some respects that the last substantial governmental research in this area was conducted in 1989.
That was now seventeen years ago. Since then, it need hardy be said, much has changed.
Business and technological practices are much different from those dismal years of two decades ago. Our consciousness of the energy crisis and of ecological issues is far better developed now than it was then. The public (electorate) is now far more aware of the issues of sustainability than they could possibly have been in the 1980s.
What’s the cost-benefit of ‘daylight saving’?
So where is the cost-benefit analysis of the different ways in wihich we might distribute the eternally pre-ordained number of daylight hours we have at our disposal, summer and winter? Common sense suggests that arranging things so there’s as much daylight as possible in the hours when most people can use it is a good start.
If anybody really knows the answer, please just let us know!
The full debate about BST is in the section of this website entitled
BST: British Summer Time & ‘Daylight Saving’ (The Clocks Go Back & Forward)…..

Read the discussion of this article which follows the book E-store…

Read the rest of this entry

Investing In The U.K.’s Big Science And Medical Research

The U.K. Science and Innovation Framework 2004-2014 has taken on new significance with the recent Budget. Scientists, economists and the regeneration arm of government need to make common cause if the proposals to reshape particle physics (PPARC), medical research (MRC) and links between business and innovation are to achieve the promise which they appear in many ways to offer.
The Government, we gather, would like to elaborate its ten year plan for science, the Science and Innovation Investment Framework 2004-2014: Next Steps, by bringing together the Particle Physics and Astronomy Council (PPARC) and the Council for the Central Laboratories of the Research Councils (CCLRC).
The proposal emerging from the Department of Trade and Industry and the Department for Education and Skills is that these two august bodies be merged as a new body, the Large Facilities Council (LFC). The LRC would have a budget of half a billion GBP a year for current CCLRC work and that part of PPARC’s work which concerns large investments. Other, grant awarding, parts of PPARC would merge with the Engineering and Physical Sciences Research Council (EPSRC).
The physicists are not happy
It would be fair to say that this proposal has not been greeted enthusiastically by everyone in the science community. To quote one astronomy blog:
This move would place astronomy and particle physics research in direct competition with the rest of the physical sciences for money. I would expect this to mean that it will be harder to get a particular research project funded, as the competition for the limited funds is greatly increased. It will also mean that the new EPSRC will have to develop a plan / road-map for the whole of engineering, physics and astronomy; a pretty huge field. Can one funding council do this alone while maintaining the breadth and depth of research in the UK?
Nor perhaps are the medics
Another of the Next Steps proposals is that the Department of Health‘s research and development budget should be merged with that of the Medical Research Council to bring together all public research in health and medicine in the U.K, with a budget of some billion GBP. Inevitably, there will be questions asked about whether this size of investment can be feasibly managed. (There are possible parallels, not least in the particle physics world, where Cern‘s much admired Large Hadron Collider (LHC) is funded by 20 European states, using the talents of 6,400 scientists from all corners of the globe.)
The wider contexts
These ideas are not, however, developing in a vacuum. Side by side with the recasting of the budgetary alignments are proposals to set targets for increasing numbers of school students, and to increase business investment and involvement in research and development. These are difficult objectives to challenge, except perhaps in the sense that ‘more not less’ might be the cry.
It’s important to acknowledge all the levels at which these various concerns and considerations apply. There are fears for vulnerable / invisible research, there are fears about the status of academic institutions and research bodies, and there are the natural fears of scientists that their jobs maybe at risk. As we know from other change initiatives, these concerns cannot simply be dismissed.
Benefits of a new kind?
We should however try to factor in a number of newer perspectives as we consider these proposals. I have argued elsewhere that support for large-scale or ‘Big Science’ in the North West of England would have been easier to secure, had there not been a stand-off between those medical scientists funded by the NHS and those funded by other bodies.
The regeneration agenda does not, as of course Gordon Brown and his colleagues would argue, stand apart from the agenda for Big Science. The real challenge, however, is to manage the necessary transitions in a way which values and promotes the knowledge economy and those who work within it, rather than leaving them behind, bewildered and resentful about the proposals which are now emerging.
Never has there been a greater need, if we are all to benefit, for the scientists, the economists, the regeneration specialists and the politicians to talk amongst themselves. This, fundamentally, is what the current consultation period on Next Steps must be about.

Big Changes In The NHS – But Where’s The Institutional Memory?

The NHS is experiencing another wave of ‘reconfiguration’, with a focus particularly on NHS Trusts and who runs them. But has there really been a shift from public sector thinking to the modern management of a complex part of the knowledge economy? On present evidence, opportunities to encapsulate hard-won insights into the organisational aspects of the health service are probably being lost.
The wholesale reconfiguration of the National Health Service, and particularly of Strategic Health Authorities and Primary Care Trusts, is now well underway. Public consultation finishes this week, and already some appointments are being progressed actively, albeit on a provisional basis.
This is probably not the place to go into the ins and outs of the basic argument – the bigger debate about the health economy is being conducted across the country and in Whitehall. What engages me more particularly is the essentially non-political issue of how much recognition is being given to the management of knowledge across this vast swathe of our public sector activity.
Is ‘institutional memory’ sufficiently acknowledged?
Whilst nealy all normal NHS employees are guaranteed jobs for a period following the intended reconfigurations, this arrangement does not extend to non-executive board positions. Indeed, the intention is now being taken forward to appoint new non-exec. posts not only through competition between existing post-holders, but with appointments being opened up to all comers – at a time when already many current non-executives will fall by the wayside anyway.
Yet non-executive directorships are the very roles which are intended to hold to account, and support, executive directors in their work. The risk is therefore that experience built up amongst non-execs. over the past few years since NHS Trusts have come into being is about to be lost, almost as soon as it has been developed.
This raises serious questions about how institutional memory and expertise amongst NHS non-executive directors is to be safeguarded. Where’s the knowledge management?
Public sector or significant knowledge economy?
Little visible effort seems so far been made to bring together in recognisable form all the aspects of high-level experience and skill which will take forward this current wave of NHS reconfiguration. Just as at last there is an emerging real understanding by non-executive directors of their crucial role, the chances are that it is to be lost again.
This is a quite separate issue from that of the general merits (or otherwise) of current moves to reshape the NHS. That debate is critical, but it is not the one we are addressing now. What we have here is a public sector health service which still sees itself as run on the basis of aspiration, rather than as a serious element of the knowledge economy, with all that implies for the management of skills, resources and the like. The preservation of institutional memory is a management, not a poltical, issue.
Introduce all the changes in structure that you wish, but alongside these must be a clear and formal recognition and management of the knowledge and skills, of themselves, within the health service. This is what modern management of complex organisations is fundamentally about, and it has to apply as much in the management of the health service as it does in any commercially-led set up.
Trying to bring (appropriate) ‘business’ attitudes into the health service is fine – though there are probably plenty of high-level people already there who have good ideas about this. But for success in the immediate future NHS organisations will need to protect and formalise their institutional knowledge right now; and the arrangements in place at present for moving on don’t make that easy.

Fire, Ice, Frost And Comets – A Lesson In Learning?

Ice & Fire (small) 06.3.2 Snow 010.jpg How we learn is always more complicated than we might imagine. The evocation of ‘fire and ice’ by both poet Robert Frost and, much later, NASA scientist Donald Brownlee, is an example to hand. Science and the arts alike depend for their impact ultimately on imagination and creativity, as well as rigour and formal insights.
Many years ago I was an American Field Service International Scholar, spending my senior high school year in Phoenix Arizona. This was a very ‘different’ experience to any I had had before or since, meeting an enormous range of people not normally encountered in suburban Birmingham England, my family home town at the time.
One of the enormous number of things I learned in Phoenix was the vast variety of interests to be found in a large American high school. And, drilling down from this, I came across a group of enthusiastic young people who Actually Read Poetry – and especially the poetry of Robert Frost, who was born on 26 March 1874 and died in 1963, not that long before I went to the States….. In a way we were perhaps a prototype Dead Poets Society, but with no sting in the tail.
A direct voice
Icy lake Then a student of science, my knowledge of poetry at the time was (and sadly remains) pretty modest, but Robert Frost’s poems fascinated me. They are direct and elemental – qualities I do not enjoy in American classical music – but also somehow quizzical, which made them very challenging in a gentle sort of way; I was never sure quite what, apart from the pastoral or earthy images, they intended to evoke. And this was especially true of Frost’s Fire and Ice, which I learnt off by heart, and can indeed still quote:

Some say the world will end in fire,
Some say in ice.
From what I’ve tasted of desire
I hold with those who favor fire.
But if it had to perish twice,
I think I know enough of hate
To say that for destruction ice
Is also great
And would suffice.
I subsequently discovered that there has been a huge amount of academic and also – engagingly – popular commentary on this poem, but at that time it simply drew me into a world away from the everyday, somewhere unknown and mysterious.
Where science meets art
Fire! (SP)16.9.05 010.jpg Given the impression Fire and Ice had on me all those years ago, I was rather startled to read the BBC News online a few days ago, reporting Dr. Donald Brownlee, as chief scientist of the Nasa Stardust mission, on evidence that comets are ‘born of fire as well as ice’.
Immediately I was transported from a murky March day in Liverpool, back some decades past to the excitement of a group of young people in a sunny classroom in Phoenix Arizona, all seeking to understand the meanings, metaphorical and material, of the complex new world into which we were about to emerge as adults.
Frost had written of fire and ice as the future destruction of world; Brownlee spoke of the birth millennia ago of physical pieces of the universe; but the elements they referred to were the same. It seems fleetingly that we are back to the phlogiston philosophers, those earlier seekers after truth, but with an up-to-the-minute twist.
All ideas are creative
Here are modern observers interpreting their experience according to their different professional disciplines, each of them evoking, for me at any rate, striking and thought-provoking images. We all carry our own paradigms as the backdrop to our understandings, but explanations are worth little without imagination to bring them alive.
I may well have been studying science when I was in the States, but Robert Frost’s poems stayed with me at least as strongly as any of the factual lessons I learned.
It would be untrue to say the science left me icy, and the poetry set my imagination on fire. But without doubt in both instances the elemental images have been retained far more strongly than the formal educational input.

Read the rest of this entry

Survival Of The Fittest In The Marketplace, But Not For Life On Earth?

Flat leaves & dew 134x124 0147aa.jpg ‘Survival of the fittest’ is often used to justify harsh business and other practices; but those who adopt this socio-economic position may also subscribe to ‘Creationist’ or ‘Intelligent Design’ notions about how life on earth has come about and diversified. This strange amalgum of beliefs arises from a lack of intellectual rigour which shows very clearly why Creationism should not become part of any serious school science curriculum.
It seems that Creationism is to be a feature of a new GCSE Biology curriculum in England. Whilst we have assurances that Creationism is not to be taught ‘as a subject’ I must admit to serious concerns about either Creationism or its close cousin, ‘Intelligent’ Design’ becoming part of the mainstream science curriculum.
General Studies can be a good place to discuss the ‘nature of science’ issues that arise from looking at Creationism or Intelligent Design, and maybe Religious Studies can offer perspectives on (non-)belief systems, but Science as a subject should probably not include subjects which are, quite simply, not subject to serious scientific scrutiny.
A conundrum of conflicting beliefs
One of the strangest things about the proponents of Intelligent Design and / or Creationism is that, for the most part, they have socio-economic beliefs which fit well within the Evolutionary Theory which they so strongly reject as an explanation of biological difference.
How can ‘survival of the fittest’ be seen to explain, and be acceptable, in terms of socio-political and economic / business affairs, but not biological ones?
Underlying this conundrum is perhaps a sense of preordination, that things are ‘given’ and cannot be changed on the whim of mere human beings; and this sense fits very helpfully into much of right-wing politics. But for those of us who respect the idea of science as a discipline and mode of knowledge, this is if anything completely the wrong way around.
Millennia to ‘change’ biology, but maybe minutes to change our own behaviour?
Living things change naturally over the millennia, ‘responding’ (i.e. surviving or not) to their contexts and inherent make up. This is a very long term and complex business.
On the other hand, human beings’ behaviour can, because we can perceive ourselves and reflect on what we do, change dramatically in the course of a single life time. ‘Intelligent Design’ is something we can all usefully engage in our own behaviour and outlook – not something which we need to dream up to ‘explain’ the amazing way in which the world as we know it has evolved over millions of years.
If survival of the fittest can be called upon by right wing thinkers to account for economic behaviour, they surely don’t have to devise other, quite undemonstrable, ‘explanations’ for the diversity of life on earth…. which leads me to wonder what less obvious reasons there may be for this strange conjunction of beliefs.

Solar Lighting Could Solve The Parks Problem

Eco- Solar (small) 06.7.15 031.jpg The debate about lighting in Liverpool’s parks continues, with strong views on both sides. One idea which resolves most of the issues raised would be solar lighting. It can be put anywhere, it’s easily maintained, it’s relatively cheap – and it has all the right ecological credentials.
Eco- Solar (azure sky) 06.7.15 030.jpg The nights are at last beginning to shorten, and we can finally think again of strolling around Sefton Park before supper.
There are big plans afoot for Sefton Park, as for several other of Liverpool’s parks, but one of the sticking points has been lighting – much of Sefton Park is unlit, and there seems little likelihood that this will change even if the ambitious renovations promised do actually come to pass.
Why no park lighting?
Several reasons have ben given for withholding lighting from large swathes of the park and its pathways (even some of the widest and most used). These include a fear that it will frighten away the bats, badgers, whatever, or that it makes unlit areas look ‘even darker and less safe’; and apparently these concerns are more compelling than the very understandable sense that a lot of people just don’t like walking in an unlit park, albeit they would like to get some exercise.
But at base I suspect that the most pressing reason for no more lighting is cost. The powers-that-be know it would be quite expensive to install and maintain, and they don’t want to ‘overburden’ the funding bids which are being developed to make our parks better and nicer places to visit.
An ecological solution
Eco- Solar (with tree & dark sky) 06.7.15 028.jpg So why can’t we bring together concerns for cost and other issues, and reach a half-way position which, to me at least, looks rather sensible?
Let’s have solar lighting.
Solar lights don’t have to be joined together with bits of cable, they don’t require electricity from a generator, they can be put anywhere (and more can be added as desired) and they don’t need time switches. Solar lights come on as it gets dark and they turn themselves off after a few hours (short stretches of light when it’s cold and only the sturdiest souls are striding out, and longer during those balmy summer evenings when everyone wants to promenade). Plus, once installed they are inexpensive, and their maintenance is easy.
And, perhaps best of all, solar lights are eco-friendly. If there’s one place in the city which needs to set an example with green credentials, surely it’s our parks?
See also: Sefton Park’s Grebes And Swans
Liverpool’s Sefton Park, Swans, Herons And Grebes

Sefton Park, Liverpool: Winter Solstice 2006
Cherry Blossom For May Day In Sefton Park, Liverpool
What Now For Liverpool’s Sefton Park?
Cherry Picking Liverpool’s Sefton Park Agenda
Liverpool’s Sefton Park Trees Under Threat – Unnecessarily?
Friends Of Sefton Park

Whatever Is The ‘Health Economy’?

The ‘health economy’ is much discussed but little defined idea. Within local health-care provision it carries an assumed status which it is perhaps now time to challenge. We don’t in everyday parlance between managers talk of an ‘education economy’; so why a ‘health economy’? Many of us would defend very strongly the concept of essential health care free at the point of delivery, but the idea of a closed specialist health economy may not be the best strategic vehicle to ensure delivery of such modern, responsive and effective health care.
There’s a fair amount of excitement around the changes in the National Health Service these days. Big shifts are about to occur in the shape, goegraphical and structural, of Primary Care Trusts, Stratgeic Health Authorities and much else. And in amongst all the other deliberations there is much reference to the ‘health economy’.
What is the ‘health economy’?
Now is probably not the best time to go into the pluses and minuses of the strategic plans for the various strands of the NHS; feelings are running high and there’s a lot to sort out yet. But it may well be a good time to ask, just what is the ‘health economy’?
This is a very particular notion, and possibly not a very helpful one. In the U.K. at least it seems mostly to refer the range of business and economic activities which fall within the scope of government-led medical attention. Nonetheless, it is by no means as easy as one might imagine to find a definition of what the health economy actually is, as opposed to simply references to it in the contexts of other health-related activity. ‘Health professions’, ‘health care’ or ‘health economics’, yes, there are many formal references and links; but ‘health economy’…. if you know of a good weblink or text book, please tell us!
A constraining concept
Perhaps it’s time to stop using this term at all. With the newer ways of delivering health care (even though this is still more likely to be ‘illness and medical care’, rather than ‘well-being and health promotion care’) the interface between different types of providers is becoming more blurred. The intention of the NHS to provide essential care free at the point of delivery remains, whoever is giving it, but the economic links are of necessity becoming far more complex.
New opportunities
There are many ways in which a more fluid concept of health-related activity might widen the scope for responsive delivery. We don’t hear about the ‘education economy’, ‘arts economy’ or ‘science economy’ as every day notions; they’re all part of, for instance, a much bigger knowledge economy. Perhaps less talk of the ‘health economy’ will open up more visible opportunities for local social enterprise and business engagement in flexible and client-responsive health care provision; and that in turn may perhaps also help local investment and regeneration in a much broader way.

Intellectuals And The ‘Post-Its Culture’

Is it true that society is more ‘anti-intellectual’ than before? How are ideas encouraged or, alternatively, left disconnected and without impact? This is a question which can be asked about the situation of both ‘thinkers’ in the accepted sense, and of people who are invited to share their views in the now well-established process of ‘community consultation’.
There was a letter recently (18.Nov.05) in New Start about ‘playing with post-its’, which I happened to re-read today. In his communication Alan Leadbetter of Stoke-on-Trent commented how the “current ‘post-it-note’ culture… encourages citizens individually to ‘have a say’,” whilst not inviting them “to take part in constructing plans, or to debate alternative plans among themselves, or to vote on them.”
This view is very much about bottom-up, grounded experience – or not, as the case may be. But by one of those strange co-incidences I also today found myself reading Frank Furedi‘s essay, Where Have All The Intellectuals Gone? Confronting 21st Century Philistinism (continuum, 2004). And yes, I was reading it in Liverpool’s very own Left Bank hangout in the city centre, the famous Everyman Bistro.
Who may produce new ideas?
Furedi’s and Leadbetter’s publications are in many ways a million miles apart; but they do have one thing in common. Both are about the formulation of new ideas, who has ‘permission’ to undertake this, and who has ownership of these ideas when they have surfaced.
Alan Leadbetter was probably thinking about people who experience significant disadvantage and rarely have an opportunity to articulate fresh ideas to any evident effect. Frank Furedi considers in his book people with wide educational and professional advantages. But the underlying connection is there.
Ideas need fertile ground
For ideas to grow, whether or not they arise from places where advantage is tangible, there must be fertile ground and space to think ‘differently’.
As yet I haven’t made my mind up whether the current perception of ‘anti-intellectualism’ is actually just that there used to be less overtly acknowledged evidence of ordinary people having important ideas at all (so that those privileged few who did ‘shine’ in this respect were more visible), or whether there is a climate now which suppresses, even more than previously, ideas which are ‘of the other’.
Either way, I do know however that fresh ideas, open to democratic interrogation, are the basis of any progressive, healthy society. There are suggestions that Philosophy become part of the general curriculm. I suppose that’s what the old ‘General Studies’, now revamped and more focused, attempted to provide. Enabling and facilitating constructive and shared ideas at every level are what it’s fundamentally about.
Maybe there’s something here we all, as community development officers, academics, teachers, politicians, media pundits, parents and citizens alike, need to think about more?